• Title/Summary/Keyword: Language Models

Search Result 872, Processing Time 0.028 seconds

Design and Implementation of Distributed QoS Management Architecture for Real-time Negotiation and Adaptation Control on CORBA Environments (CORBA 환경에서 실시간 협약 및 작응 제어를 위한 분사 QoS 관리 구조의 설계 및 구현)

  • Lee, Won-Jung;Shin, Chang-Sun;Jeong, Chang-Won;Joo, Su-Chong
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.1C
    • /
    • pp.21-35
    • /
    • 2002
  • Nowadays, in accordance with increasing expectations of multimedia stream service on the internet, a lot of distributed applications are being required and developed. But the models of the existing systems have the problems that cannot support the extensibility and the reusability, when the QoS relating functions are being developed as an integrated modules which are suited on the centralized controlled specific-purpose application services. To cope with these problems, it is suggested in this paper to a distributed QoS management system on CORBA, an object-oriented middleware compliance. This systems we suggested can provides not only for efficient control of resources, various service QoS, and QoS control functions as the existing functions, but also QoS control real-time negotiation and dynamic adaptation in addition. This system consists of QoS Control Management Module(QoS CMM) in client side and QoS Management Module(QoS MM) in server side, respectively. These distributed modules are interfacing with each other via CORBA on different systems for distributed QoS management while serving distributed streaming applications. In phase of design of our system, we use UML(Unified Modeling Language) for designing each component in modules, their method calls and various detailed functions for controlling QoS of stream services. For implementation of our system, we used OrbixWeb 3.1c following CORBA specification on Solaris 2.5/2.7, Java language, Java Media Framework API 2.0 beta2, Mini-SQL 1.0.16 and the multimedia equipments, such as SunVideoPlus/Sun Video capture board and Sun Camera. Finally, we showed a numerical data controlled by real-time negotiation and adaptation procedures based on QoS map information to GUIs on client and server dynamically, while our distributed QoS management system is executing a given streaming service.

The Design and Implementation of e-BCOS for e-Business Component System (e-비즈니스 컴포넌트 시스템 설계 및 구현)

  • Choi, Ha-Jung;Kim, Haeng-Kon
    • The KIPS Transactions:PartD
    • /
    • v.10D no.1
    • /
    • pp.85-100
    • /
    • 2003
  • Today's computing system has expanded its application to business trade and distributed work transactions using the Internet. As the demand for more flexible, adaptable, extensible, and robust web-based enterprise, these application development has been gradually expanded based on reusable, independent, and portable components. Component Based Development (CBD) works by developing and evolving software from selected reusable software components and then assembling them within appropriate software architecture. However, it requires an increase in cost to build new components as well as the necessary effort to develop of the business requirement these components. Standardized component models are required as well from the perspective of systems in order to support rapid and exact component information transmission on the web. In this paper, we describe the e-Business Component Development with agent for rapid application development on the web that correspond to the demands of users in the business domain. We design and implement the specifications of e-business components by combining these demands. In order to improve the agent register and retrieval, we propose the intelligent search and register agents, which can conduct more precise searching and specializing for components. The system enables the locating of user's frequently used components through an agent involving register and retrieval, as well as rapid procedures for registers The e-BCOS (e-Business Component System) is the agent system for the user to register distributed components and to search for components Information. The e-BCOS increases reusability through the e-business component development of distributed components in the business domain. For the share and delivery, specification with XML is acceptable to user's variable order e-BCOS Includes the effective investment, timeliness, reliability, efficiency, and maintenance effort by with agent.

Analysis of Optimum Design of Stepped Bar Horn for 20kHz Metal Ultrasonic Welding (20kHz 급 금속 초음파 융착용 스텝형 바 혼의 최적설계)

  • Kim, Jisun;Kim, Jaewoong;Kim, In-ju;Seo, Joowhan
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.20 no.12
    • /
    • pp.94-101
    • /
    • 2019
  • In this study, the FEM technique was applied to design the shape of the horn that transmits ultrasonic vibration energy to the base material, and the shape of the welding horn with a one-wavelength bar shape used in the 20kHz region was designed. The shape design of the horn was performed by applying the rod longitudinal vibration theory to Ansys APDL (Ansys Parametric Design Language). Twenty-five models were designed using the ratio of the area of the input and output surfaces of the vibration and the length of the horn to derive the appropriate horn shape. The horn was designed with a total length of 130mm, a step length of 65mm, and an output area of 28.79mm. The horn was fabricated using the optimized dimensions, and the vibration and displacement characteristics of the horn were evaluated using the measurement system. Finally, a uniform longitudinal step horn was designed, and more than 97.4% of the uniformity of the tip was secured. The amplitude ratio of the optimized horn was improved by 51%.

An RDB to RDF Mapping System Considering Semantic Relations of RDB Components (관계형 데이터베이스 구성 요소의 의미 관계를 고려한 RDB to RDF 매핑 시스템)

  • Sung, Hajung;Gim, Jangwon;Lee, Sukhoon;Baik, Doo-Kwon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.1
    • /
    • pp.19-30
    • /
    • 2014
  • For the expansion of the Semantic Web, studies in converting the data stored in the relational database into the ontology are actively in process. Such studies mainly use an RDB to RDF mapping model, the model to map relational database components to RDF components. However, pre-proposed mapping models have got different expression modes and these damage the accessibility and reusability of the users. As a consequence, the necessity of the standardized mapping language was raised and the W3C suggested the R2RML as the standard mapping language for the RDB to RDF model. The R2RML has a characteristic that converts only the relational database schema data to RDF. For the same reasons above, the ontology about the relation data between table name and column name of the relational database cannot be added. In this paper, we propose an RDB to RDF mapping system considering semantic relations of RDB components in order to solve the above issue. The proposed system generates the mapping data by adding the RDFS attribute data into the schema data defined by the R2RML in the relational database. This mapping data converts the data stored in the relational database into RDF which includes the RDFS attribute data. In this paper, we implement the proposed system as a Java-based prototype, perform the experiment which converts the data stored in the relational database into RDF for the comparison evaluation purpose and compare the results against D2RQ, RDBToOnto and Morph. The proposed system expresses semantic relations which has richer converted ontology than any other studies and shows the best performance in data conversion time.

A Comic Facial Expression Method for Intelligent Avatar Communications in the Internet Cyberspace (인터넷 가상공간에서 지적 아바타 통신을 위한 코믹한 얼굴 표정의 생성법)

  • 이용후;김상운;청목유직
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.40 no.1
    • /
    • pp.59-73
    • /
    • 2003
  • As a means of overcoming the linguistic barrier between different languages in the Internet, a new sign-language communication system with CG animation techniques has been developed and proposed. In the system, the joint angles of the arms and the hands corresponding to the gesture as a non-verbal communication tool have been considered. The emotional expression, however, could as play also an important role in communicating each other. Especially, a comic expression is more efficient than real facial expression, and the movements of the cheeks and the jaws are more important AU's than those of the eyebrow, eye, mouth etc. Therefore, in this paper, we designed a 3D emotion editor using 2D model, and we extract AU's (called as PAU, here) which play a principal function in expressing emotions. We also proposed a method of generating the universal emotional expression with Avatar models which have different vertex structures. Here, we employed a method of dynamically adjusting the AU movements according to emotional intensities. The proposed system is implemented with Visual C++ and Open Inventor on windows platforms. Experimental results show a possibility that the system could be used as a non-verbal communication means to overcome the linguistic barrier.

Deep learning-based Multilingual Sentimental Analysis using English Review Data (영어 리뷰데이터를 이용한 딥러닝 기반 다국어 감성분석)

  • Sung, Jae-Kyung;Kim, Yung Bok;Kim, Yong-Guk
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.3
    • /
    • pp.9-15
    • /
    • 2019
  • Large global online shopping malls, such as Amazon, offer services in English or in the language of a country when their products are sold. Since many customers purchase products based on the product reviews, the shopping malls actively utilize the sentimental analysis technique in judging preference of each product using the large amount of review data that the customer has written. And the result of such analysis can be used for the marketing to look the potential shoppers. However, it is difficult to apply this English-based semantic analysis system to different languages used around the world. In this study, more than 500,000 data from Amazon fine food reviews was used for training a deep learning based system. First, sentiment analysis evaluation experiments were carried out with three models of English test data. Secondly, the same data was translated into seven languages (Korean, Japanese, Chinese, Vietnamese, French, German and English) and then the similar experiments were done. The result suggests that although the accuracy of the sentimental analysis was 2.77% lower than the average of the seven countries (91.59%) compared to the English (94.35%), it is believed that the results of the experiment can be used for practical applications.

Semantic Computing-based Dynamic Job Scheduling Model and Simulation (시멘틱 컴퓨팅 기반의 동적 작업 스케줄링 모델 및 시뮬레이션)

  • Noh, Chang-Hyeon;Jang, Sung-Ho;Kim, Tae-Young;Lee, Jong-Sik
    • Journal of the Korea Society for Simulation
    • /
    • v.18 no.2
    • /
    • pp.29-38
    • /
    • 2009
  • In the computing environment with heterogeneous resources, a job scheduling model is necessary for effective resource utilization and high-speed data processing. And, the job scheduling model has to cope with a dynamic change in the condition of resources. There have been lots of researches on resource estimation methods and heuristic algorithms about how to distribute and allocate jobs to heterogeneous resources. But, existing researches have a weakness for system compatibility and scalability because they do not support the standard language. Also, they are impossible to process jobs effectively and deal with a variety of computing situations in which the condition of resources is dynamically changed in real-time. In order to solve the problems of existing researches, this paper proposes a semantic computing-based dynamic job scheduling model that defines various knowledge-based rules for job scheduling methods adaptable to changes in resource condition and allocate a job to the best suited resource through inference. This paper also constructs a resource ontology to manage information about heterogeneous resources without difficulty as using the OWL, the standard ontology language established by W3C. Experimental results shows that the proposed scheduling model outperforms existing scheduling models, in terms of throughput, job loss, and turn around time.

Building robust Korean speech recognition model by fine-tuning large pretrained model (대형 사전훈련 모델의 파인튜닝을 통한 강건한 한국어 음성인식 모델 구축)

  • Changhan Oh;Cheongbin Kim;Kiyoung Park
    • Phonetics and Speech Sciences
    • /
    • v.15 no.3
    • /
    • pp.75-82
    • /
    • 2023
  • Automatic speech recognition (ASR) has been revolutionized with deep learning-based approaches, among which self-supervised learning methods have proven to be particularly effective. In this study, we aim to enhance the performance of OpenAI's Whisper model, a multilingual ASR system on the Korean language. Whisper was pretrained on a large corpus (around 680,000 hours) of web speech data and has demonstrated strong recognition performance for major languages. However, it faces challenges in recognizing languages such as Korean, which is not major language while training. We address this issue by fine-tuning the Whisper model with an additional dataset comprising about 1,000 hours of Korean speech. We also compare its performance against a Transformer model that was trained from scratch using the same dataset. Our results indicate that fine-tuning the Whisper model significantly improved its Korean speech recognition capabilities in terms of character error rate (CER). Specifically, the performance improved with increasing model size. However, the Whisper model's performance on English deteriorated post fine-tuning, emphasizing the need for further research to develop robust multilingual models. Our study demonstrates the potential of utilizing a fine-tuned Whisper model for Korean ASR applications. Future work will focus on multilingual recognition and optimization for real-time inference.

Automated Story Generation with Image Captions and Recursiva Calls (이미지 캡션 및 재귀호출을 통한 스토리 생성 방법)

  • Isle Jeon;Dongha Jo;Mikyeong Moon
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.24 no.1
    • /
    • pp.42-50
    • /
    • 2023
  • The development of technology has achieved digital innovation throughout the media industry, including production techniques and editing technologies, and has brought diversity in the form of consumer viewing through the OTT service and streaming era. The convergence of big data and deep learning networks automatically generated text in format such as news articles, novels, and scripts, but there were insufficient studies that reflected the author's intention and generated story with contextually smooth. In this paper, we describe the flow of pictures in the storyboard with image caption generation techniques, and the automatic generation of story-tailored scenarios through language models. Image caption using CNN and Attention Mechanism, we generate sentences describing pictures on the storyboard, and input the generated sentences into the artificial intelligence natural language processing model KoGPT-2 in order to automatically generate scenarios that meet the planning intention. Through this paper, the author's intention and story customized scenarios are created in large quantities to alleviate the pain of content creation, and artificial intelligence participates in the overall process of digital content production to activate media intelligence.

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.