• Title/Summary/Keyword: Big data Processing

Search Result 1,063, Processing Time 0.031 seconds

An Investigation on Digital Humanities Research Trend by Analyzing the Papers of Digital Humanities Conferences (디지털 인문학 연구 동향 분석 - Digital Humanities 학술대회 논문을 중심으로 -)

  • Chung, EunKyung
    • Journal of the Korean Society for Library and Information Science
    • /
    • v.55 no.1
    • /
    • pp.393-413
    • /
    • 2021
  • Digital humanities, which creates new and innovative knowledge through the combination of digital information technology and humanities research problems, can be seen as a representative multidisciplinary field of study. To investigate the intellectual structure of the digital humanities field, a network analysis of authors and keywords co-word was performed on a total of 441 papers in the last two years (2019, 2020) at the Digital Humanities Conference. As the results of the author and keyword analysis show, we can find out the active activities of Europe, North America, and Japanese and Chinese authors in East Asia. Through the co-author network, 11 dis-connected sub-networks are identified, which can be seen as a result of closed co-authoring activities. Through keyword analysis, 16 sub-subject areas are identified, which are machine learning, pedagogy, metadata, topic modeling, stylometry, cultural heritage, network, digital archive, natural language processing, digital library, twitter, drama, big data, neural network, virtual reality, and ethics. This results imply that a diver variety of digital information technologies are playing a major role in the digital humanities. In addition, keywords with high frequency can be classified into humanities-based keywords, digital information technology-based keywords, and convergence keywords. The dynamics of the growth and development of digital humanities can represented in these combinations of keywords.

A Study on the Current State of the Library's AI Service and the Service Provision Plan (도서관의 인공지능(AI) 서비스 현황 및 서비스 제공 방안에 관한 연구)

  • Kwak, Woojung;Noh, Younghee
    • Journal of Korean Library and Information Science Society
    • /
    • v.52 no.1
    • /
    • pp.155-178
    • /
    • 2021
  • In the era of the 4th industrial revolution, public libraries need a strategy for promoting intelligent library services in order to actively respond to changes in the external environment such as artificial intelligence. Therefore, in this study, based on the concept of artificial intelligence and analysis of domestic and foreign artificial intelligence related trends, policies, and cases, we proposed the future direction of introduction and development of artificial intelligence services in the library. Currently, the library operates a reference information service that automatically provides answers through the introduction of artificial intelligence technologies such as deep learning and natural language processing, and develops a big data-based AI book recommendation and automatic book inspection system to increase business utilization and provide customized services for users. Has been provided. In the field of companies and industries, regardless of domestic and overseas, we are developing and servicing technologies based on autonomous driving using artificial intelligence, personal customization, etc., and providing optimal results by self-learning information using deep learning. It is developed in the form of an equation. Accordingly, in the future, libraries will utilize artificial intelligence to recommend personalized books based on the user's usage records, recommend reading and culture programs, and introduce real-time delivery services through transport methods such as autonomous drones and cars in the case of book delivery service. Service development should be promoted.

CoAID+ : COVID-19 News Cascade Dataset for Social Context Based Fake News Detection (CoAID+ : 소셜 컨텍스트 기반 가짜뉴스 탐지를 위한 COVID-19 뉴스 파급 데이터)

  • Han, Soeun;Kang, Yoonsuk;Ko, Yunyong;Ahn, Jeewon;Kim, Yushim;Oh, Seongsoo;Park, Heejin;Kim, Sang-Wook
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.4
    • /
    • pp.149-156
    • /
    • 2022
  • In the current COVID-19 pandemic, fake news and misinformation related to COVID-19 have been causing serious confusion in our society. To accurately detect such fake news, social context-based methods have been widely studied in the literature. They detect fake news based on the social context that indicates how a news article is propagated over social media (e.g., Twitter). Most existing COVID-19 related datasets gathered for fake news detection, however, contain only the news content information, but not its social context information. In this case, the social context-based detection methods cannot be applied, which could be a big obstacle in the fake news detection research. To address this issue, in this work, we collect from Twitter the social context information based on CoAID, which is a COVID-19 news content dataset built for fake news detection, thereby building CoAID+ that includes both the news content information and its social context information. The CoAID+ dataset can be utilized in a variety of methods for social context-based fake news detection, thus would help revitalize the fake news detection research area. Finally, through a comprehensive analysis of the CoAID+ dataset in various perspectives, we present some interesting features capable of differentiating real and fake news.

Voice Interactions with A. I. Agent : Analysis of Domestic and Overseas IT Companies (A.I.에이전트와의 보이스 인터랙션 : 국내외 IT회사 사례연구)

  • Lee, Seo-Young
    • Journal of Korea Entertainment Industry Association
    • /
    • v.15 no.4
    • /
    • pp.15-29
    • /
    • 2021
  • Many countries and companies are pursuing and developing Artificial intelligence as it is the core technology of the 4th industrial revolution. Global IT companies such as Apple, Microsoft, Amazon, Google and Samsung have all released their own AI assistant hardware products, hoping to increase customer loyalty and capture market share. Competition within the industry for AI agent is intense. AI assistant products that command the biggest market shares and customer loyalty have a higher chance of becoming the industry standard. This study analyzed the current status of major overseas and domestic IT companies in the field of artificial intelligence, and suggested future strategic directions for voice UI technology development and user satisfaction. In terms of B2B technology, it is recommended that IT companies use cloud computing to store big data, innovative artificial intelligence technologies and natural language technologies. Offering voice recognition technologies on the cloud enables smaller companies to take advantage of such technologies at considerably less expense. Companies also consider using GPT-3(Generative Pre-trained Transformer 3) an open source artificial intelligence language processing software that can generate very natural human-like interactions and high levels of user satisfaction. There is a need to increase usefulness and usability to enhance user satisfaction. This study has practical and theoretical implications for industry and academia.

The Application Methods of FarmMap Reading in Agricultural Land Using Deep Learning (딥러닝을 이용한 농경지 팜맵 판독 적용 방안)

  • Wee Seong Seung;Jung Nam Su;Lee Won Suk;Shin Yong Tae
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.2
    • /
    • pp.77-82
    • /
    • 2023
  • The Ministry of Agriculture, Food and Rural Affairs established the FarmMap, an digital map of agricultural land. In this study, using deep learning, we suggest the application of farm map reading to farmland such as paddy fields, fields, ginseng, fruit trees, facilities, and uncultivated land. The farm map is used as spatial information for planting status and drone operation by digitizing agricultural land in the real world using aerial and satellite images. A reading manual has been prepared and updated every year by demarcating the boundaries of agricultural land and reading the attributes. Human reading of agricultural land differs depending on reading ability and experience, and reading errors are difficult to verify in reality because of budget limitations. The farmmap has location information and class information of the corresponding object in the image of 5 types of farmland properties, so the suitable AI technique was tested with ResNet50, an instance segmentation model. The results of attribute reading of agricultural land using deep learning and attribute reading by humans were compared. If technology is developed by focusing on attribute reading that shows different results in the future, it is expected that it will play a big role in reducing attribute errors and improving the accuracy of digital map of agricultural land.

GIS Based Distributed Flood Damage Assessment (GIS기반의 분포형 홍수피해산정 기법)

  • Yi, Choong Sung;Choi, Seung An;Shim, Myung Pil;Kim, Hung Soo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.26 no.3B
    • /
    • pp.301-310
    • /
    • 2006
  • Typically, we needs enormous national budget for the flood control project and so the project usually has big influence on the national economy. Therefore, the reliable estimation of flood damage is the key issue for the economic analysis of the flood control project. This study aims to provide a GIS based technique for distributed flood damage estimation. We consider two aspects of engineering and economic sides, which are the inundation analysis and MD-FDA (Multi-Dimensional Flood Damage Analysis), for the flood damage assessment. We propose the analysis framework and data processing using GIS for assessing flood damages. The proposed methodology is applied to the flood control channel project for flood disaster prevention in Mokgamcheon/Dorimcheon streams and this study presents the detailed GIS database and the assessment results of flood damages. This study may have the worth in improving practical usability of MD-FDA and also providing research direction for combining economic side with the engineering aspect. Also this distributed technique will help decision-making in evaluating the feasibility of flood damage reduction programs for structural and nonstructural measures.

An Analysis on Characteristics of Turbulence Energy Dissipation Rate from Comparison of Wind Profiler and Rawinsonde (연직바람관측장비와 레윈존데의 비교를 통한 난류 에너지 감소률의 특성 분석)

  • Kang, Woo Kyeong;Moon, Yun Seob;Jung, Ok Jin
    • Journal of the Korean earth science society
    • /
    • v.37 no.7
    • /
    • pp.448-464
    • /
    • 2016
  • The purpose of this study is to optimize the parameters related to consensus coherency within the PCL 1300, the operating program of wind profiler, from a validation of wind data between rawinsonde and wind profiler at Chupungryeong ($36^{\circ}13^{\prime}$, $127^{\circ}59^{\prime}$) site in Korea. It is then to analyze the diurnal and seasonal characteristics of the turbulence energy dissipation rate (${\varepsilon}$) in clear and rainy days from March 2009 to February 2010. In comparison of the wind data between wind profiler and rawinsonde during April 22-23, 2010, it was shown in a big error more than $10ms^{-1}$ over the height of 3,000 meters in the zonal (u) and meridional (v) wind components. When removing more than $10ms^{-1}$ in each wind speed difference of u an v components between the two instruments, the correlation coefficients of these wind components were 0.92 and 0.88, respectively, and the root mean square errors were 3.07 and $1.06ms^{-1}$. Based on these results, when the data processing time and the minimum available data within the PCL 1300 program were adjusted as 30 minutes and 60%, respectively, the bias errors were small. In addition, as a result of an analysis of sensitivity to consensus coherency of u and v components within the PCL1300 program, u components were underestimated in radial coherency, instantaneous and winbarbs coherency, whereas v components were overestimated. Finally by optimizing parameters of the PCL1300 program, the diurnal and seasonal means of ${\varepsilon}$ at each height were higher in rainy days than those in clear days because of increasing in the vertical wind speed due to upward and downward motions. The mean ${\varepsilon}$ for clear and rainy days in winter was lower than those of other seasons, due to stronger horizontal wind speed in winter than those in other seasons. Consequently, when the turbulence energy dissipation rates in the vertical wind speed of more than ${\pm}10cm\;s^{-1}$ were excluded for clear and rainy days, the mean ${\varepsilon}$ in rainy days was 6-7 times higher than that in clear days, but when considering them, it was 4-5 times higher.

The Stock Portfolio Recommendation System based on the Correlation between the Stock Message Boards and the Stock Market (인터넷 주식 토론방 게시물과 주식시장의 상관관계 분석을 통한 투자 종목 선정 시스템)

  • Lee, Yun-Jung;Kim, Gun-Woo;Woo, Gyun
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.10
    • /
    • pp.441-450
    • /
    • 2014
  • The stock market is constantly changing and sometimes the stock prices unaccountably plummet or surge. So, the stock market is recognized as a complex system and the change on the stock prices is unpredictable. Recently, many researchers try to understand the stock market as the network among individual stocks and to find a clue about the change of the stock prices from big data being created in real time from Internet. We focus on the correlation between the stock prices and the human interactions in Internet especially in the stock message boards. To uncover this correlation, we collected and investigated the articles concerning with 57 target companies, members of KOSPI200. From the analysis result, we found that there is no significant correlation between the stock prices and the article volume, but the strength of correlation between the article volume and the stock prices is relevant to the stock return. We propose a new method for recommending stock portfolio base on the result of our analysis. According to the simulated investment test using the article data from the stock message boards in 'Daum' portal site, the returns of our portfolio is about 1.55% per month, which is about 0.72% and 1.21% higher than that of the Markowitz's efficient portfolio and that of the KOSPI average respectively. Also, the case using the data from 'Naver' portal site, the stock returns of our proposed portfolio is about 0.90%, which is 0.35%, 0.40%, and 0.58% higher than those of our previous portfolio, Markowitz's efficient portfolio, and KOSPI average respectively. This study presents that collective human behavior on Internet stock message board can be much helpful to understand the stock market and the correlation between the stock price and the collective human behavior can be used to invest in stocks.

A Mobile Landmarks Guide : Outdoor Augmented Reality based on LOD and Contextual Device (모바일 랜드마크 가이드 : LOD와 문맥적 장치 기반의 실외 증강현실)

  • Zhao, Bi-Cheng;Rosli, Ahmad Nurzid;Jang, Chol-Hee;Lee, Kee-Sung;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.1
    • /
    • pp.1-21
    • /
    • 2012
  • In recent years, mobile phone has experienced an extremely fast evolution. It is equipped with high-quality color displays, high resolution cameras, and real-time accelerated 3D graphics. In addition, some other features are includes GPS sensor and Digital Compass, etc. This evolution advent significantly helps the application developers to use the power of smart-phones, to create a rich environment that offers a wide range of services and exciting possibilities. To date mobile AR in outdoor research there are many popular location-based AR services, such Layar and Wikitude. These systems have big limitation the AR contents hardly overlaid on the real target. Another research is context-based AR services using image recognition and tracking. The AR contents are precisely overlaid on the real target. But the real-time performance is restricted by the retrieval time and hardly implement in large scale area. In our work, we exploit to combine advantages of location-based AR with context-based AR. The system can easily find out surrounding landmarks first and then do the recognition and tracking with them. The proposed system mainly consists of two major parts-landmark browsing module and annotation module. In landmark browsing module, user can view an augmented virtual information (information media), such as text, picture and video on their smart-phone viewfinder, when they pointing out their smart-phone to a certain building or landmark. For this, landmark recognition technique is applied in this work. SURF point-based features are used in the matching process due to their robustness. To ensure the image retrieval and matching processes is fast enough for real time tracking, we exploit the contextual device (GPS and digital compass) information. This is necessary to select the nearest and pointed orientation landmarks from the database. The queried image is only matched with this selected data. Therefore, the speed for matching will be significantly increased. Secondly is the annotation module. Instead of viewing only the augmented information media, user can create virtual annotation based on linked data. Having to know a full knowledge about the landmark, are not necessary required. They can simply look for the appropriate topic by searching it with a keyword in linked data. With this, it helps the system to find out target URI in order to generate correct AR contents. On the other hand, in order to recognize target landmarks, images of selected building or landmark are captured from different angle and distance. This procedure looks like a similar processing of building a connection between the real building and the virtual information existed in the Linked Open Data. In our experiments, search range in the database is reduced by clustering images into groups according to their coordinates. A Grid-base clustering method and user location information are used to restrict the retrieval range. Comparing the existed research using cluster and GPS information the retrieval time is around 70~80ms. Experiment results show our approach the retrieval time reduces to around 18~20ms in average. Therefore the totally processing time is reduced from 490~540ms to 438~480ms. The performance improvement will be more obvious when the database growing. It demonstrates the proposed system is efficient and robust in many cases.

Open Digital Textbook for Smart Education (스마트교육을 위한 오픈 디지털교과서)

  • Koo, Young-Il;Park, Choong-Shik
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.177-189
    • /
    • 2013
  • In Smart Education, the roles of digital textbook is very important as face-to-face media to learners. The standardization of digital textbook will promote the industrialization of digital textbook for contents providers and distributers as well as learner and instructors. In this study, the following three objectives-oriented digital textbooks are looking for ways to standardize. (1) digital textbooks should undertake the role of the media for blended learning which supports on-off classes, should be operating on common EPUB viewer without special dedicated viewer, should utilize the existing framework of the e-learning learning contents and learning management. The reason to consider the EPUB as the standard for digital textbooks is that digital textbooks don't need to specify antoher standard for the form of books, and can take advantage od industrial base with EPUB standards-rich content and distribution structure (2) digital textbooks should provide a low-cost open market service that are currently available as the standard open software (3) To provide appropriate learning feedback information to students, digital textbooks should provide a foundation which accumulates and manages all the learning activity information according to standard infrastructure for educational Big Data processing. In this study, the digital textbook in a smart education environment was referred to open digital textbook. The components of open digital textbooks service framework are (1) digital textbook terminals such as smart pad, smart TVs, smart phones, PC, etc., (2) digital textbooks platform to show and perform digital contents on digital textbook terminals, (3) learning contents repository, which exist on the cloud, maintains accredited learning, (4) App Store providing and distributing secondary learning contents and learning tools by learning contents developing companies, and (5) LMS as a learning support/management tool which on-site class teacher use for creating classroom instruction materials. In addition, locating all of the hardware and software implement a smart education service within the cloud must have take advantage of the cloud computing for efficient management and reducing expense. The open digital textbooks of smart education is consdered as providing e-book style interface of LMS to learners. In open digital textbooks, the representation of text, image, audio, video, equations, etc. is basic function. But painting, writing, problem solving, etc are beyond the capabilities of a simple e-book. The Communication of teacher-to-student, learner-to-learnert, tems-to-team is required by using the open digital textbook. To represent student demographics, portfolio information, and class information, the standard used in e-learning is desirable. To process learner tracking information about the activities of the learner for LMS(Learning Management System), open digital textbook must have the recording function and the commnincating function with LMS. DRM is a function for protecting various copyright. Currently DRMs of e-boook are controlled by the corresponding book viewer. If open digital textbook admitt DRM that is used in a variety of different DRM standards of various e-book viewer, the implementation of redundant features can be avoided. Security/privacy functions are required to protect information about the study or instruction from a third party UDL (Universal Design for Learning) is learning support function for those with disabilities have difficulty in learning courses. The open digital textbook, which is based on E-book standard EPUB 3.0, must (1) record the learning activity log information, and (2) communicate with the server to support the learning activity. While the recording function and the communication function, which is not determined on current standards, is implemented as a JavaScript and is utilized in the current EPUB 3.0 viewer, ths strategy of proposing such recording and communication functions as the next generation of e-book standard, or special standard (EPUB 3.0 for education) is needed. Future research in this study will implement open source program with the proposed open digital textbook standard and present a new educational services including Big Data analysis.