• Title/Summary/Keyword: Intelligent Data Analysis

Search Result 1,456, Processing Time 0.03 seconds

Analyzing the Effects of Low Emission Bus Zones Using Bus Information System Data (버스정보시스템 데이터를 활용한 Low Emission Bus Zone 도입의 탄소배출 저감 효과 분석)

  • Hye Inn Song;Kangwon Shin
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.22 no.6
    • /
    • pp.196-207
    • /
    • 2023
  • As part of measures to address the climate crisis, buses are also being converted to electric and hydrogen buses. Local authorities need to prioritize carbon emissions when allocating newly introduced and converted electric and hydrogen buses, and as a method, consider the introduction of Low Emission Bus Zones (LEBZ) to propose the reduction of pollution from specific links. To introduce LEBZ, it is necessary to compare the carbon emissions before and after its implementation, yet there is a shortage of studies that focus solely on buses or analyze the effects of introducing LEBZ to specific links. In this paper, we utilized bus information system data to calculate and compare the effects of introducing LEBZ to bus priority lanes in Jeju. We categorized scenarios into five groups, with scenarios 1 through 4 involving the introduction of LEBZ, and scenario 5 designating cases where LEBZ was not introduced. Comparative results confirmed that in scenarios with LEBZ introduction, the reduction per km reached a maximum of 0.097t per km, whereas in cases without LEBZ, it amounted to 0.022t per km, demonstrating higher efficiency. It underscores the significance of conducting carbon emission calculations and comparing the effects of LEBZ introduction using bus information system data, which can be directly applied by local authorities to make informed and rational decisions.

Tour-based Personalized Trip Analysis and Calibration Method for Activity-based Traffic Demand Modelling (활동기반 교통수요 모델링을 위한 투어기반 통행분석 및 보정방안)

  • Yegi Yoo;Heechan Kang;Seungmo Yoo;Taeho Oh
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.22 no.6
    • /
    • pp.32-48
    • /
    • 2023
  • Autonomous driving technology is shaping the future of personalized travel, encouraging personalized travel, and traffic impact could be influenced by individualized travel behavior during the transition of driving entity from human to machine. In order to evaluate traffic impact, it is necessary to estimate the total number of trips based on an understanding of individual travel characteristics. The Activity-based model(ABM), which allows for the reflection of individual travel characteristics, deals with all travel sequences of an individual. Understanding the relationship between travel and travel must be important for assessing traffic impact using ABM. However, the ABM has a limitation in the data hunger model. It is difficult to adjust in the actual demand forecasting. Therefore, we utilized a Tour-based model that can explain the relationship between travels based on household travel survey data instead. After that, vehicle registration and population data were used for correction. The result showed that, compared to the KTDB one, the traffic generation exhibited a 13% increase in total trips and approximately 9% reduction in working trips, valid within an acceptable margin of error. As a result, it can be used as a generation correction method based on Tour, which can reflect individual travel characteristics, prior to building an activity-based model to predict demand due to the introduction of autonomous vehicles in terms of road operation, which is the ultimate goal of this study.

A Study on the Capacity Review of One-lane Hi-pass Lanes on Highways : Focusing on Using Bootstrapping Techniques (고속도로 단차로 하이패스차로 용량 검토에 관한 연구 : 부트스트랩 기법 활용 중심으로)

  • Bosung Kim;Donghee Han
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.23 no.3
    • /
    • pp.1-16
    • /
    • 2024
  • In the present highway design guidelines suggest that the capacity of one-lane hi-pass lanes is 2,000 veh/h for mainline toll plaza and 1,700 veh/h for interchange toll plaza. However, in a study conducted in early 2010, capacity of the mainline toll plaza was presented with 1,476 veh/h/ln to 1,665 veh/h/ln, and capacity of the interchange toll plaza was presented as 1,443 veh/h/ln. Accordingly, this study examined the feasibility of the capacity of the currently proposed highway one-lane hi-pass lane. Based on the 2021 individual vehicle passing data collected from the one-lane hi-pass gantry, the speed-traffic volume relationship graph and headway were used to calculate and compare capacity. In addition, the bootstrapping technique was introduced to utilize the headway and new processing methods for collected data were reviewed. As a result of the analysis, the one-lane hi-pass capacity could be estimated at 1,700 veh/h/ln for the interchange toll plaza, and at least 1,700 veh/h/ln for the mainline toll plaza. In addition, by using the bootstrap technique when using headway data, it was possible to present an estimated capacity similar to the observed capacity.

Recommender system using BERT sentiment analysis (BERT 기반 감성분석을 이용한 추천시스템)

  • Park, Ho-yeon;Kim, Kyoung-jae
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.2
    • /
    • pp.1-15
    • /
    • 2021
  • If it is difficult for us to make decisions, we ask for advice from friends or people around us. When we decide to buy products online, we read anonymous reviews and buy them. With the advent of the Data-driven era, IT technology's development is spilling out many data from individuals to objects. Companies or individuals have accumulated, processed, and analyzed such a large amount of data that they can now make decisions or execute directly using data that used to depend on experts. Nowadays, the recommender system plays a vital role in determining the user's preferences to purchase goods and uses a recommender system to induce clicks on web services (Facebook, Amazon, Netflix, Youtube). For example, Youtube's recommender system, which is used by 1 billion people worldwide every month, includes videos that users like, "like" and videos they watched. Recommended system research is deeply linked to practical business. Therefore, many researchers are interested in building better solutions. Recommender systems use the information obtained from their users to generate recommendations because the development of the provided recommender systems requires information on items that are likely to be preferred by the user. We began to trust patterns and rules derived from data rather than empirical intuition through the recommender systems. The capacity and development of data have led machine learning to develop deep learning. However, such recommender systems are not all solutions. Proceeding with the recommender systems, there should be no scarcity in all data and a sufficient amount. Also, it requires detailed information about the individual. The recommender systems work correctly when these conditions operate. The recommender systems become a complex problem for both consumers and sellers when the interaction log is insufficient. Because the seller's perspective needs to make recommendations at a personal level to the consumer and receive appropriate recommendations with reliable data from the consumer's perspective. In this paper, to improve the accuracy problem for "appropriate recommendation" to consumers, the recommender systems are proposed in combination with context-based deep learning. This research is to combine user-based data to create hybrid Recommender Systems. The hybrid approach developed is not a collaborative type of Recommender Systems, but a collaborative extension that integrates user data with deep learning. Customer review data were used for the data set. Consumers buy products in online shopping malls and then evaluate product reviews. Rating reviews are based on reviews from buyers who have already purchased, giving users confidence before purchasing the product. However, the recommendation system mainly uses scores or ratings rather than reviews to suggest items purchased by many users. In fact, consumer reviews include product opinions and user sentiment that will be spent on evaluation. By incorporating these parts into the study, this paper aims to improve the recommendation system. This study is an algorithm used when individuals have difficulty in selecting an item. Consumer reviews and record patterns made it possible to rely on recommendations appropriately. The algorithm implements a recommendation system through collaborative filtering. This study's predictive accuracy is measured by Root Mean Squared Error (RMSE) and Mean Absolute Error (MAE). Netflix is strategically using the referral system in its programs through competitions that reduce RMSE every year, making fair use of predictive accuracy. Research on hybrid recommender systems combining the NLP approach for personalization recommender systems, deep learning base, etc. has been increasing. Among NLP studies, sentiment analysis began to take shape in the mid-2000s as user review data increased. Sentiment analysis is a text classification task based on machine learning. The machine learning-based sentiment analysis has a disadvantage in that it is difficult to identify the review's information expression because it is challenging to consider the text's characteristics. In this study, we propose a deep learning recommender system that utilizes BERT's sentiment analysis by minimizing the disadvantages of machine learning. This study offers a deep learning recommender system that uses BERT's sentiment analysis by reducing the disadvantages of machine learning. The comparison model was performed through a recommender system based on Naive-CF(collaborative filtering), SVD(singular value decomposition)-CF, MF(matrix factorization)-CF, BPR-MF(Bayesian personalized ranking matrix factorization)-CF, LSTM, CNN-LSTM, GRU(Gated Recurrent Units). As a result of the experiment, the recommender system based on BERT was the best.

A MVC Framework for Visualizing Text Data (텍스트 데이터 시각화를 위한 MVC 프레임워크)

  • Choi, Kwang Sun;Jeong, Kyo Sung;Kim, Soo Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.39-58
    • /
    • 2014
  • As the importance of big data and related technologies continues to grow in the industry, it has become highlighted to visualize results of processing and analyzing big data. Visualization of data delivers people effectiveness and clarity for understanding the result of analyzing. By the way, visualization has a role as the GUI (Graphical User Interface) that supports communications between people and analysis systems. Usually to make development and maintenance easier, these GUI parts should be loosely coupled from the parts of processing and analyzing data. And also to implement a loosely coupled architecture, it is necessary to adopt design patterns such as MVC (Model-View-Controller) which is designed for minimizing coupling between UI part and data processing part. On the other hand, big data can be classified as structured data and unstructured data. The visualization of structured data is relatively easy to unstructured data. For all that, as it has been spread out that the people utilize and analyze unstructured data, they usually develop the visualization system only for each project to overcome the limitation traditional visualization system for structured data. Furthermore, for text data which covers a huge part of unstructured data, visualization of data is more difficult. It results from the complexity of technology for analyzing text data as like linguistic analysis, text mining, social network analysis, and so on. And also those technologies are not standardized. This situation makes it more difficult to reuse the visualization system of a project to other projects. We assume that the reason is lack of commonality design of visualization system considering to expanse it to other system. In our research, we suggest a common information model for visualizing text data and propose a comprehensive and reusable framework, TexVizu, for visualizing text data. At first, we survey representative researches in text visualization era. And also we identify common elements for text visualization and common patterns among various cases of its. And then we review and analyze elements and patterns with three different viewpoints as structural viewpoint, interactive viewpoint, and semantic viewpoint. And then we design an integrated model of text data which represent elements for visualization. The structural viewpoint is for identifying structural element from various text documents as like title, author, body, and so on. The interactive viewpoint is for identifying the types of relations and interactions between text documents as like post, comment, reply and so on. The semantic viewpoint is for identifying semantic elements which extracted from analyzing text data linguistically and are represented as tags for classifying types of entity as like people, place or location, time, event and so on. After then we extract and choose common requirements for visualizing text data. The requirements are categorized as four types which are structure information, content information, relation information, trend information. Each type of requirements comprised with required visualization techniques, data and goal (what to know). These requirements are common and key requirement for design a framework which keep that a visualization system are loosely coupled from data processing or analyzing system. Finally we designed a common text visualization framework, TexVizu which is reusable and expansible for various visualization projects by collaborating with various Text Data Loader and Analytical Text Data Visualizer via common interfaces as like ITextDataLoader and IATDProvider. And also TexVisu is comprised with Analytical Text Data Model, Analytical Text Data Storage and Analytical Text Data Controller. In this framework, external components are the specifications of required interfaces for collaborating with this framework. As an experiment, we also adopt this framework into two text visualization systems as like a social opinion mining system and an online news analysis system.

Analysis of Artificial Intelligence's Technology Innovation and Diffusion Pattern: Focusing on USPTO Patent Data (인공지능의 기술 혁신 및 확산 패턴 분석: USPTO 특허 데이터를 중심으로)

  • Baek, Seoin;Lee, Hyunjin;Kim, Heetae
    • The Journal of the Korea Contents Association
    • /
    • v.20 no.4
    • /
    • pp.86-98
    • /
    • 2020
  • The artificial intelligence (AI) is a technology that will lead the future connective and intelligent era by combining with almost all industries in manufacturing and service industry. Although Korea is one of the world's leading artificial intelligence group with the United States, Japan, and Germany, but its competitiveness in terms of artificial intelligence patent is relatively low compared to others. Therefore, it is necessary to carry out quantitative analysis of artificial intelligence patents in various aspects in order to examine national competitiveness, major industries and future development directions in artificial intelligence technology. In this study, we use the IPC technology classification code to estimate the overall life cycle and the speed of development of the artificial intelligence technology. We collected patents related to artificial intelligence from 2008 to 2018, and analyze patent trends through one-dimensional statistical analysis, two-dimensional statistical analysis and network analysis. We expect that the technological trends of the artificial intelligence industry discovered from this study will be exploited to the strategies of the artificial intelligence technology and the policy making of the government.

An Implementation of Optimum Tender Price Automatic Calculation System using Statistical Analysis Technique (통계분석 기법을 이용한 최적의 투찰가 자동 산출 시스템의 구현)

  • Kim, Bong-Hyun;Lee, Se-Hwan;Cho, Dong-Uk
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.33 no.11B
    • /
    • pp.1013-1019
    • /
    • 2008
  • Recently, various information and data are efficiently used by the rapid growth of its Internet in our real life. But, users have spent lots of time finding necessary information for the increased amounts of information. To solve this problem, it can be provided the speed, accuracy of information search with development of intelligent search engines, agent system etc. In this paper, we propose the method of getting the best tender price in the analysis of the construction bid information that needs its professionalism by on the purpose to maximize users' satisfaction. Of course, if it is not under the unit of a results in the future, we put target of this paper on part to heighten supreme successful bid success rate. Therefore, this paper embodies offered system of web based on producing tender price of most suitable through techniques to produce tender price about successful bid that compare with bidder fare by statistical analysis incidental and value approaching successful bidder fare by frequency analysis method.

Security Analysis of the Whirlpool Hash Function in the Cloud of Things

  • Li, Wei;Gao, Zhiyong;Gu, Dawu;Ge, Chenyu;Liao, Linfeng;Zhou, Zhihong;Liu, Ya;Liu, Zhiqiang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.1
    • /
    • pp.536-551
    • /
    • 2017
  • With the advancement and deployment of leading-edge telecommunication technologies for sensing and collecting, computing related information, Cloud of Things (CoTs) has emerged as a typical application platform that is envisioned to revolutionize the daily activities of human society, such as intelligent transportation, modern logistics, food safety, environmental monitoring, etc. To avoid any possible malicious attack and resource abuse, employing hash functions is widely recognized as one of the most effective approaches for CoTs to achieve message integrity and data authentication. The Whirlpool hash function has served as part of the joint ISO/IEC 10118-3 International Standard by the International Organization for Standardization (ISO) and the International Electrotechnical Commission (IEC). In this paper, we propose an effective differential fault analysis on Whirlpool in the byte-oriented random fault model. The mathematical analysis and experimental results show that 8 random faults on average are required to obtain the current 512-bit message input of whirlpool and the secret key of HMAC-Whirlpool. Our work demonstrates that Whirlpool and HMAC-Whirlpool are both vulnerable to the single byte differential fault analysis. It provides a new reference for the security analysis of the same structure of the hash functions in the CoTs.

Class Analysis Method Using Video Synchronization Algorithm (동영상 동기화 알고리즘을 이용한 수업 분석 방법)

  • Kwon, Ohsung
    • Journal of The Korean Association of Information Education
    • /
    • v.19 no.4
    • /
    • pp.441-448
    • /
    • 2015
  • This paper describes about a software implementation for class analysis and quantization based on our video synchronization method. We proposed a new indexing method, synchronization strategies, and data structure for our analyzer implementation. We implemented a class video analyzer using intelligent multimedia technologies which can play class video selectively. Our proposed method analyzes class videos depending on the time schedule composed of introduction, development and summary stages. We apply our analysis filters to the class videos in the predefined regular intervals. We experimented on the synchronization performance of our proposed method and software. In the experimental, we could demonstrate the effectiveness and practicality of our class analyzing method within the margin of error.

A Study on Forecasting Accuracy Improvement of Case Based Reasoning Approach Using Fuzzy Relation (퍼지 관계를 활용한 사례기반추론 예측 정확성 향상에 관한 연구)

  • Lee, In-Ho;Shin, Kyung-Shik
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.4
    • /
    • pp.67-84
    • /
    • 2010
  • In terms of business, forecasting is a work of what is expected to happen in the future to make managerial decisions and plans. Therefore, the accurate forecasting is very important for major managerial decision making and is the basis for making various strategies of business. But it is very difficult to make an unbiased and consistent estimate because of uncertainty and complexity in the future business environment. That is why we should use scientific forecasting model to support business decision making, and make an effort to minimize the model's forecasting error which is difference between observation and estimator. Nevertheless, minimizing the error is not an easy task. Case-based reasoning is a problem solving method that utilizes the past similar case to solve the current problem. To build the successful case-based reasoning models, retrieving the case not only the most similar case but also the most relevant case is very important. To retrieve the similar and relevant case from past cases, the measurement of similarities between cases is an important key factor. Especially, if the cases contain symbolic data, it is more difficult to measure the distances. The purpose of this study is to improve the forecasting accuracy of case-based reasoning approach using fuzzy relation and composition. Especially, two methods are adopted to measure the similarity between cases containing symbolic data. One is to deduct the similarity matrix following binary logic(the judgment of sameness between two symbolic data), the other is to deduct the similarity matrix following fuzzy relation and composition. This study is conducted in the following order; data gathering and preprocessing, model building and analysis, validation analysis, conclusion. First, in the progress of data gathering and preprocessing we collect data set including categorical dependent variables. Also, the data set gathered is cross-section data and independent variables of the data set include several qualitative variables expressed symbolic data. The research data consists of many financial ratios and the corresponding bond ratings of Korean companies. The ratings we employ in this study cover all bonds rated by one of the bond rating agencies in Korea. Our total sample includes 1,816 companies whose commercial papers have been rated in the period 1997~2000. Credit grades are defined as outputs and classified into 5 rating categories(A1, A2, A3, B, C) according to credit levels. Second, in the progress of model building and analysis we deduct the similarity matrix following binary logic and fuzzy composition to measure the similarity between cases containing symbolic data. In this process, the used types of fuzzy composition are max-min, max-product, max-average. And then, the analysis is carried out by case-based reasoning approach with the deducted similarity matrix. Third, in the progress of validation analysis we verify the validation of model through McNemar test based on hit ratio. Finally, we draw a conclusion from the study. As a result, the similarity measuring method using fuzzy relation and composition shows good forecasting performance compared to the similarity measuring method using binary logic for similarity measurement between two symbolic data. But the results of the analysis are not statistically significant in forecasting performance among the types of fuzzy composition. The contributions of this study are as follows. We propose another methodology that fuzzy relation and fuzzy composition could be applied for the similarity measurement between two symbolic data. That is the most important factor to build case-based reasoning model.