• Title/Summary/Keyword: Large language models

Search Result 155, Processing Time 0.043 seconds

Building Light Weight CORBA Based Middleware for the CAN Bus Systems

  • Hong, Seongsoo
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.3 no.3
    • /
    • pp.181-189
    • /
    • 2001
  • The software components of embedded control systems get extremely complex as they are designed into distributed systems get extremely complex as they are designed into distributed systems consisting of a large number of inexpensive microcontrollers interconnected by low-bandwidth real-time networks such as the controller area network (CAN). While recently emerging middleware technologies such as CORBA and DCOM address the complexity of distributed programming, they cannot be directly applied to distributed control system design due to their excessive resource demand and inadequate communication models. In this paper, we propose a CORBA-based middleware design for CAN-based distributed embedded control systems. Our design goal is to minimize its resource need and make it support group communication without losing the IDL (interface definition language) level compliance to the OMG standards. To achieve this, we develop a transport protocol on the CAN and a group communication scheme based on the well-known publisher/subscriber model. The protocol effectively realizes subject-based addressing and supports anonymous publisher/subscriber communication. We also customize the method invocation and message passing protocol, referred to as the general inter-ORB protocol (GIOP), of CORBA so that CORBA method invocations are efficiently serviced on a low-bandwidth network such as the CAN. This customization includes packed data encoding and variable-length integer encoding for compact representation of IDL data types. We have implemented our CORBA-based middleware on the mArx real-time operating system we have developed at Seoul National University. Our experiments clearly demonstrate that it is feasible to use CORBA in developing distributed embedded control systems possessing severe resource limitations. Our design clearly demonstrates that it is feasible to use a CORBA-based middleware in developing distributed embedded systems on real-time networks possessing severe resource limitations.

  • PDF

The Parallel Corpus Approach to Building the Syntactic Tree Transfer Set in the English-to- Vietnamese Machine Translation

  • Dien Dinh;Ngan Thuy;Quang Xuan;Nam Chi
    • Proceedings of the IEEK Conference
    • /
    • summer
    • /
    • pp.382-386
    • /
    • 2004
  • Recently, with the machine learning trend, most of the machine translation systems on over the world use two syntax tree sets of two relevant languages to learn syntactic tree transfer rules. However, for the English-Vietnamese language pair, this approach is impossible because until now we have not had a Vietnamese syntactic tree set which is correspondent to English one. Building of a very large correspondent Vietnamese syntactic tree set (thousands of trees) requires so much work and take the investment of specialists in linguistics. To take advantage from our available English-Vietnamese Corpus (EVC) which was tagged in word alignment, we choose the SITG (Stochastic Inversion Transduction Grammar) model to construct English- Vietnamese syntactic tree sets automatically. This model is used to parse two languages at the same time and then carry out the syntactic tree transfer. This English-Vietnamese bilingual syntactic tree set is the basic training data to carry out transferring automatically from English syntactic trees to Vietnamese ones by machine learning models. We tested the syntax analysis by comparing over 10,000 sentences in the amount of 500,000 sentences of our English-Vietnamese bilingual corpus and first stage got encouraging result $(analyzed\;about\;80\%)[5].$ We have made use the TBL algorithm (Transformation Based Learning) to carry out automatic transformations from English syntactic trees to Vietnamese ones based on that parallel syntactic tree transfer set[6].

  • PDF

An XML-based Digital Mock-Up System for Heterogeneous Multi-CAD Assembly (XML을 이용한 이기종 CAD 조립체 DMU시스템의 설계)

  • Song, In-Ho;Chung, Sung-Chong
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.31 no.6 s.261
    • /
    • pp.635-643
    • /
    • 2007
  • As many engineers and technicians are involved in the design process of large scale and/or complex products, there are a lot of miss matches and interferences due to designers' faults and several kinds of CAD systems. Recently, CAD systems are applied to verify and check the assembly process. Digital Mock-Up(DMU) system, a tool to build a virtual mock-up in the design stage, has been used to prevent the interferences and miss matches during precision design processes. Using the virtual assembly tool, engineers are able to design precision and interference free parts without physical mock-ups. Instead of a single CAD source, several CAD systems are used to design a complex product. Several organizations are involved in the distributed design environment for heterogeneous multi-CAD assembly. XML and the lightweight CAD file are proposed for the multi-CAD assembly. XML data contains hierarchy of the heterogenenous multi-CAD assembly. STEP PDM schema and STEP ISO 10303-28 formations are applied to construct the XML data. The lightweight CAD file produced from various CAD files through ACIS kernel and InterOp not only contains mesn, B-Rep and topological data, but also is used to visualize CAD data and to verify dimensions. Developed system is executed on the desktop computers. It does not require commercial CAD systems to visualize 3D assembly data. Real-time interference and fitness checks, dimensional verification, and design and assembly verification are performed on the developed system. Assembly of heterogeneous models for a car is conducted to verify the effectiveness of the developed DMU system on the Internet.

Design and Implementation of Peer-to-Peer Electronic Commerce Systems based on the File Sharing Method between Users (이용자간 파일공유방식에 기반한 P2P 전자상거래 시스템 설계 및 구현)

  • Kim Chang-Su;Seo Young-Suk
    • The Journal of Information Systems
    • /
    • v.15 no.1
    • /
    • pp.1-20
    • /
    • 2006
  • Peer-to-peer systems (P2P) are rapidly growing in importance on the Internet environment, quickly extending the range of their usage. However, peer-to-peer systems have not been widely applied in electronic commerce because they have not been established as an appropriate business model. Therefore, we firstly review the previous research relevant to peer-to-peer systems, and then analyze the business models for P2P systems presented by previous researchers. Furthermore, this study categorizes major issues in terms of the technical and business model aspects. On the basis of these reviews, we develop P2P electronic commerce systems based on the file sharing method between users, focusing on user interface friendliness. A developed P2P electronic commerce systems are programmed by using the C# based on the Microsoft.net solution. A database is implemented using the MSSQL2000. A main application technology is designed that P2P electronic commerce systems make it possible. for user to extend into BtoB Solution by using WSDL (Web Services Description Language), UDDI (Universal Description, Discovery, and Integration) and the XML that is a document for users. User interface is made as form of Internet messenger for a user's convenience and is possible to develop into a commodity transaction system based on XML. In this study, it is possible for the P2P electronic commerce system to have extended application to fields such as Internet shopping mall and property transaction in a nonprofit organization, a public institution and a large scale nonprofit institution that have a similar structure as compared with a structure of a nonprofit educational institution.

  • PDF

Integrated Structural Design Operation by Process Decomposition and Parallelization (프로세스 분할 병행에 의한 통합 구조설계 운용)

  • Hwang, Jin-Ha;Park, Jong-Hoi
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.21 no.1
    • /
    • pp.113-124
    • /
    • 2008
  • Distributed operation of overall structural design process, by which product optimization and process parallelization are simultaneously implemented, is presented in this paper. The database-interacted hybrid method, which selectively takes the accustomed procedure of the conventional method in the framework of the optimal design, is utilized here. The staged application of design constraints reduces the computational burden for large complex optimization problems. Two kinds of numeric and graphic processes are simultaneously implemented by concurrent engineering approach in the distributed environment of PC networks. The former is based on finite element optimization method and the latter is represented by AutoCAD using AutoLISP programming language. Numerical computation and database interaction on servers and graphic works on independent clients are communicated through message passing. The numerical experiments for some steel truss models show the validity and usability of the method. This study has sufficient adaptability and expandability, in that it is based on general methodologies and industry standard platforms.

Automated Story Generation with Image Captions and Recursiva Calls (이미지 캡션 및 재귀호출을 통한 스토리 생성 방법)

  • Isle Jeon;Dongha Jo;Mikyeong Moon
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.24 no.1
    • /
    • pp.42-50
    • /
    • 2023
  • The development of technology has achieved digital innovation throughout the media industry, including production techniques and editing technologies, and has brought diversity in the form of consumer viewing through the OTT service and streaming era. The convergence of big data and deep learning networks automatically generated text in format such as news articles, novels, and scripts, but there were insufficient studies that reflected the author's intention and generated story with contextually smooth. In this paper, we describe the flow of pictures in the storyboard with image caption generation techniques, and the automatic generation of story-tailored scenarios through language models. Image caption using CNN and Attention Mechanism, we generate sentences describing pictures on the storyboard, and input the generated sentences into the artificial intelligence natural language processing model KoGPT-2 in order to automatically generate scenarios that meet the planning intention. Through this paper, the author's intention and story customized scenarios are created in large quantities to alleviate the pain of content creation, and artificial intelligence participates in the overall process of digital content production to activate media intelligence.

KNU Korean Sentiment Lexicon: Bi-LSTM-based Method for Building a Korean Sentiment Lexicon (Bi-LSTM 기반의 한국어 감성사전 구축 방안)

  • Park, Sang-Min;Na, Chul-Won;Choi, Min-Seong;Lee, Da-Hee;On, Byung-Won
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.219-240
    • /
    • 2018
  • Sentiment analysis, which is one of the text mining techniques, is a method for extracting subjective content embedded in text documents. Recently, the sentiment analysis methods have been widely used in many fields. As good examples, data-driven surveys are based on analyzing the subjectivity of text data posted by users and market researches are conducted by analyzing users' review posts to quantify users' reputation on a target product. The basic method of sentiment analysis is to use sentiment dictionary (or lexicon), a list of sentiment vocabularies with positive, neutral, or negative semantics. In general, the meaning of many sentiment words is likely to be different across domains. For example, a sentiment word, 'sad' indicates negative meaning in many fields but a movie. In order to perform accurate sentiment analysis, we need to build the sentiment dictionary for a given domain. However, such a method of building the sentiment lexicon is time-consuming and various sentiment vocabularies are not included without the use of general-purpose sentiment lexicon. In order to address this problem, several studies have been carried out to construct the sentiment lexicon suitable for a specific domain based on 'OPEN HANGUL' and 'SentiWordNet', which are general-purpose sentiment lexicons. However, OPEN HANGUL is no longer being serviced and SentiWordNet does not work well because of language difference in the process of converting Korean word into English word. There are restrictions on the use of such general-purpose sentiment lexicons as seed data for building the sentiment lexicon for a specific domain. In this article, we construct 'KNU Korean Sentiment Lexicon (KNU-KSL)', a new general-purpose Korean sentiment dictionary that is more advanced than existing general-purpose lexicons. The proposed dictionary, which is a list of domain-independent sentiment words such as 'thank you', 'worthy', and 'impressed', is built to quickly construct the sentiment dictionary for a target domain. Especially, it constructs sentiment vocabularies by analyzing the glosses contained in Standard Korean Language Dictionary (SKLD) by the following procedures: First, we propose a sentiment classification model based on Bidirectional Long Short-Term Memory (Bi-LSTM). Second, the proposed deep learning model automatically classifies each of glosses to either positive or negative meaning. Third, positive words and phrases are extracted from the glosses classified as positive meaning, while negative words and phrases are extracted from the glosses classified as negative meaning. Our experimental results show that the average accuracy of the proposed sentiment classification model is up to 89.45%. In addition, the sentiment dictionary is more extended using various external sources including SentiWordNet, SenticNet, Emotional Verbs, and Sentiment Lexicon 0603. Furthermore, we add sentiment information about frequently used coined words and emoticons that are used mainly on the Web. The KNU-KSL contains a total of 14,843 sentiment vocabularies, each of which is one of 1-grams, 2-grams, phrases, and sentence patterns. Unlike existing sentiment dictionaries, it is composed of words that are not affected by particular domains. The recent trend on sentiment analysis is to use deep learning technique without sentiment dictionaries. The importance of developing sentiment dictionaries is declined gradually. However, one of recent studies shows that the words in the sentiment dictionary can be used as features of deep learning models, resulting in the sentiment analysis performed with higher accuracy (Teng, Z., 2016). This result indicates that the sentiment dictionary is used not only for sentiment analysis but also as features of deep learning models for improving accuracy. The proposed dictionary can be used as a basic data for constructing the sentiment lexicon of a particular domain and as features of deep learning models. It is also useful to automatically and quickly build large training sets for deep learning models.

Deep Learning OCR based document processing platform and its application in financial domain (금융 특화 딥러닝 광학문자인식 기반 문서 처리 플랫폼 구축 및 금융권 내 활용)

  • Dongyoung Kim;Doohyung Kim;Myungsung Kwak;Hyunsoo Son;Dongwon Sohn;Mingi Lim;Yeji Shin;Hyeonjung Lee;Chandong Park;Mihyang Kim;Dongwon Choi
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.1
    • /
    • pp.143-174
    • /
    • 2023
  • With the development of deep learning technologies, Artificial Intelligence powered Optical Character Recognition (AI-OCR) has evolved to read multiple languages from various forms of images accurately. For the financial industry, where a large number of diverse documents are processed through manpower, the potential for using AI-OCR is great. In this study, we present a configuration and a design of an AI-OCR modality for use in the financial industry and discuss the platform construction with application cases. Since the use of financial domain data is prohibited under the Personal Information Protection Act, we developed a deep learning-based data generation approach and used it to train the AI-OCR models. The AI-OCR models are trained for image preprocessing, text recognition, and language processing and are configured as a microservice architected platform to process a broad variety of documents. We have demonstrated the AI-OCR platform by applying it to financial domain tasks of document sorting, document verification, and typing assistance The demonstrations confirm the increasing work efficiency and conveniences.

Comparison of Deep Learning Frameworks: About Theano, Tensorflow, and Cognitive Toolkit (딥러닝 프레임워크의 비교: 티아노, 텐서플로, CNTK를 중심으로)

  • Chung, Yeojin;Ahn, SungMahn;Yang, Jiheon;Lee, Jaejoon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.1-17
    • /
    • 2017
  • The deep learning framework is software designed to help develop deep learning models. Some of its important functions include "automatic differentiation" and "utilization of GPU". The list of popular deep learning framework includes Caffe (BVLC) and Theano (University of Montreal). And recently, Microsoft's deep learning framework, Microsoft Cognitive Toolkit, was released as open-source license, following Google's Tensorflow a year earlier. The early deep learning frameworks have been developed mainly for research at universities. Beginning with the inception of Tensorflow, however, it seems that companies such as Microsoft and Facebook have started to join the competition of framework development. Given the trend, Google and other companies are expected to continue investing in the deep learning framework to bring forward the initiative in the artificial intelligence business. From this point of view, we think it is a good time to compare some of deep learning frameworks. So we compare three deep learning frameworks which can be used as a Python library. Those are Google's Tensorflow, Microsoft's CNTK, and Theano which is sort of a predecessor of the preceding two. The most common and important function of deep learning frameworks is the ability to perform automatic differentiation. Basically all the mathematical expressions of deep learning models can be represented as computational graphs, which consist of nodes and edges. Partial derivatives on each edge of a computational graph can then be obtained. With the partial derivatives, we can let software compute differentiation of any node with respect to any variable by utilizing chain rule of Calculus. First of all, the convenience of coding is in the order of CNTK, Tensorflow, and Theano. The criterion is simply based on the lengths of the codes and the learning curve and the ease of coding are not the main concern. According to the criteria, Theano was the most difficult to implement with, and CNTK and Tensorflow were somewhat easier. With Tensorflow, we need to define weight variables and biases explicitly. The reason that CNTK and Tensorflow are easier to implement with is that those frameworks provide us with more abstraction than Theano. We, however, need to mention that low-level coding is not always bad. It gives us flexibility of coding. With the low-level coding such as in Theano, we can implement and test any new deep learning models or any new search methods that we can think of. The assessment of the execution speed of each framework is that there is not meaningful difference. According to the experiment, execution speeds of Theano and Tensorflow are very similar, although the experiment was limited to a CNN model. In the case of CNTK, the experimental environment was not maintained as the same. The code written in CNTK has to be run in PC environment without GPU where codes execute as much as 50 times slower than with GPU. But we concluded that the difference of execution speed was within the range of variation caused by the different hardware setup. In this study, we compared three types of deep learning framework: Theano, Tensorflow, and CNTK. According to Wikipedia, there are 12 available deep learning frameworks. And 15 different attributes differentiate each framework. Some of the important attributes would include interface language (Python, C ++, Java, etc.) and the availability of libraries on various deep learning models such as CNN, RNN, DBN, and etc. And if a user implements a large scale deep learning model, it will also be important to support multiple GPU or multiple servers. Also, if you are learning the deep learning model, it would also be important if there are enough examples and references.

Analysis of the Impact of Generative AI based on Crunchbase: Before and After the Emergence of ChatGPT (Crunchbase를 바탕으로 한 Generative AI 영향 분석: ChatGPT 등장 전·후를 중심으로)

  • Nayun Kim;Youngjung Geum
    • Asia-Pacific Journal of Business Venturing and Entrepreneurship
    • /
    • v.19 no.3
    • /
    • pp.53-68
    • /
    • 2024
  • Generative AI is receiving a lot of attention around the world, and ways to effectively utilize it in the business environment are being explored. In particular, since the public release of the ChatGPT service, which applies the GPT-3.5 model, a large language model developed by OpenAI, it has attracted more attention and has had a significant impact on the entire industry. This study focuses on the emergence of Generative AI, especially ChatGPT, which applies OpenAI's GPT-3.5 model, to investigate its impact on the startup industry and compare the changes that occurred before and after its emergence. This study aims to shed light on the actual application and impact of generative AI in the business environment by examining in detail how generative AI is being used in the startup industry and analyzing the impact of ChatGPT's emergence on the industry. To this end, we collected company information of generative AI-related startups that appeared before and after the ChatGPT announcement and analyzed changes in industry, business content, and investment information. Through keyword analysis, topic modeling, and network analysis, we identified trends in the startup industry and how the introduction of generative AI has revolutionized the startup industry. As a result of the study, we found that the number of startups related to Generative AI has increased since the emergence of ChatGPT, and in particular, the total and average amount of funding for Generative AI-related startups has increased significantly. We also found that various industries are attempting to apply Generative AI technology, and the development of services and products such as enterprise applications and SaaS using Generative AI has been actively promoted, influencing the emergence of new business models. The findings of this study confirm the impact of Generative AI on the startup industry and contribute to our understanding of how the emergence of this innovative new technology can change the business ecosystem.

  • PDF