• Title/Summary/Keyword: Topic Generation

Search Result 159, Processing Time 0.026 seconds

Flow solutions around rectangular cylinders: The question of spatial discretization

  • Corsini, Roberto;Angeli, Diego;Stalio, Enrico;Chibbaro, Sergio;Cimarelli, Andrea
    • Wind and Structures
    • /
    • v.34 no.1
    • /
    • pp.151-159
    • /
    • 2022
  • The aerodynamics of blunt bodies with separation at the sharp corner of the leading edge and reattachment on the body side are particularly important in civil engineering applications. In recent years, a number of experimental and numerical studies have become available on the aerodynamics of a rectangular cylinder with chord-to-thickness ratio equal to 5 (BARC). Despite the interest in the topic, a widely accepted set of guidelines for grid generation about these blunt bodies is still missing. In this work a new, well resolved Direct Numerical Simulation (DNS) around the BARC body at Re=3000 is presented and its results compared to previous DNSs of the same case but with different numerical approaches and mesh. Despite the simulations use different numerical approaches, mesh and domain dimensions, the main discrepancies are ascribed to the different grid spacings employed. While a more rigorous analysis is envisaged, where the order of accuracy of the schemes are kept the same while grid spacings are varied alternately along each spatial direction, this represents a first attempt in the study of the influence of spatial resolution in the Direct Numerical Simulation of flows around elongated rectangular cylinders with sharp corners.

A Mobile Platform System based on SBC (SBC 기반 차세대 이동형 단말기 개발)

  • Lee, Seung-Ik
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.14 no.4
    • /
    • pp.30-36
    • /
    • 2009
  • In this paper, we develop the next generation mobile system based on SBC(Server Based Computing). SBC is one of the popular topic and only one server system is used as multiple personal systems with a terminal, a monitor and a keyboard. The advantages of this system are easy-upgrade, low costs and convenience for each user. But it is difficult to design and manufacture the multimedia streaming system and devices control system for external devices. In this paper, we represent and develops the Mobile system based on SBC and the test performances shows that our system has good performances of multimedia system and security for internet.

Twitter Issue Tracking System by Topic Modeling Techniques (토픽 모델링을 이용한 트위터 이슈 트래킹 시스템)

  • Bae, Jung-Hwan;Han, Nam-Gi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.109-122
    • /
    • 2014
  • People are nowadays creating a tremendous amount of data on Social Network Service (SNS). In particular, the incorporation of SNS into mobile devices has resulted in massive amounts of data generation, thereby greatly influencing society. This is an unmatched phenomenon in history, and now we live in the Age of Big Data. SNS Data is defined as a condition of Big Data where the amount of data (volume), data input and output speeds (velocity), and the variety of data types (variety) are satisfied. If someone intends to discover the trend of an issue in SNS Big Data, this information can be used as a new important source for the creation of new values because this information covers the whole of society. In this study, a Twitter Issue Tracking System (TITS) is designed and established to meet the needs of analyzing SNS Big Data. TITS extracts issues from Twitter texts and visualizes them on the web. The proposed system provides the following four functions: (1) Provide the topic keyword set that corresponds to daily ranking; (2) Visualize the daily time series graph of a topic for the duration of a month; (3) Provide the importance of a topic through a treemap based on the score system and frequency; (4) Visualize the daily time-series graph of keywords by searching the keyword; The present study analyzes the Big Data generated by SNS in real time. SNS Big Data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. In addition, such analysis requires the latest big data technology to process rapidly a large amount of real-time data, such as the Hadoop distributed system or NoSQL, which is an alternative to relational database. We built TITS based on Hadoop to optimize the processing of big data because Hadoop is designed to scale up from single node computing to thousands of machines. Furthermore, we use MongoDB, which is classified as a NoSQL database. In addition, MongoDB is an open source platform, document-oriented database that provides high performance, high availability, and automatic scaling. Unlike existing relational database, there are no schema or tables with MongoDB, and its most important goal is that of data accessibility and data processing performance. In the Age of Big Data, the visualization of Big Data is more attractive to the Big Data community because it helps analysts to examine such data easily and clearly. Therefore, TITS uses the d3.js library as a visualization tool. This library is designed for the purpose of creating Data Driven Documents that bind document object model (DOM) and any data; the interaction between data is easy and useful for managing real-time data stream with smooth animation. In addition, TITS uses a bootstrap made of pre-configured plug-in style sheets and JavaScript libraries to build a web system. The TITS Graphical User Interface (GUI) is designed using these libraries, and it is capable of detecting issues on Twitter in an easy and intuitive manner. The proposed work demonstrates the superiority of our issue detection techniques by matching detected issues with corresponding online news articles. The contributions of the present study are threefold. First, we suggest an alternative approach to real-time big data analysis, which has become an extremely important issue. Second, we apply a topic modeling technique that is used in various research areas, including Library and Information Science (LIS). Based on this, we can confirm the utility of storytelling and time series analysis. Third, we develop a web-based system, and make the system available for the real-time discovery of topics. The present study conducted experiments with nearly 150 million tweets in Korea during March 2013.

Subject-Balanced Intelligent Text Summarization Scheme (주제 균형 지능형 텍스트 요약 기법)

  • Yun, Yeoil;Ko, Eunjung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.141-166
    • /
    • 2019
  • Recently, channels like social media and SNS create enormous amount of data. In all kinds of data, portions of unstructured data which represented as text data has increased geometrically. But there are some difficulties to check all text data, so it is important to access those data rapidly and grasp key points of text. Due to needs of efficient understanding, many studies about text summarization for handling and using tremendous amounts of text data have been proposed. Especially, a lot of summarization methods using machine learning and artificial intelligence algorithms have been proposed lately to generate summary objectively and effectively which called "automatic summarization". However almost text summarization methods proposed up to date construct summary focused on frequency of contents in original documents. Those summaries have a limitation for contain small-weight subjects that mentioned less in original text. If summaries include contents with only major subject, bias occurs and it causes loss of information so that it is hard to ascertain every subject documents have. To avoid those bias, it is possible to summarize in point of balance between topics document have so all subject in document can be ascertained, but still unbalance of distribution between those subjects remains. To retain balance of subjects in summary, it is necessary to consider proportion of every subject documents originally have and also allocate the portion of subjects equally so that even sentences of minor subjects can be included in summary sufficiently. In this study, we propose "subject-balanced" text summarization method that procure balance between all subjects and minimize omission of low-frequency subjects. For subject-balanced summary, we use two concept of summary evaluation metrics "completeness" and "succinctness". Completeness is the feature that summary should include contents of original documents fully and succinctness means summary has minimum duplication with contents in itself. Proposed method has 3-phases for summarization. First phase is constructing subject term dictionaries. Topic modeling is used for calculating topic-term weight which indicates degrees that each terms are related to each topic. From derived weight, it is possible to figure out highly related terms for every topic and subjects of documents can be found from various topic composed similar meaning terms. And then, few terms are selected which represent subject well. In this method, it is called "seed terms". However, those terms are too small to explain each subject enough, so sufficient similar terms with seed terms are needed for well-constructed subject dictionary. Word2Vec is used for word expansion, finds similar terms with seed terms. Word vectors are created after Word2Vec modeling, and from those vectors, similarity between all terms can be derived by using cosine-similarity. Higher cosine similarity between two terms calculated, higher relationship between two terms defined. So terms that have high similarity values with seed terms for each subjects are selected and filtering those expanded terms subject dictionary is finally constructed. Next phase is allocating subjects to every sentences which original documents have. To grasp contents of all sentences first, frequency analysis is conducted with specific terms that subject dictionaries compose. TF-IDF weight of each subjects are calculated after frequency analysis, and it is possible to figure out how much sentences are explaining about each subjects. However, TF-IDF weight has limitation that the weight can be increased infinitely, so by normalizing TF-IDF weights for every subject sentences have, all values are changed to 0 to 1 values. Then allocating subject for every sentences with maximum TF-IDF weight between all subjects, sentence group are constructed for each subjects finally. Last phase is summary generation parts. Sen2Vec is used to figure out similarity between subject-sentences, and similarity matrix can be formed. By repetitive sentences selecting, it is possible to generate summary that include contents of original documents fully and minimize duplication in summary itself. For evaluation of proposed method, 50,000 reviews of TripAdvisor are used for constructing subject dictionaries and 23,087 reviews are used for generating summary. Also comparison between proposed method summary and frequency-based summary is performed and as a result, it is verified that summary from proposed method can retain balance of all subject more which documents originally have.

Finding Weighted Sequential Patterns over Data Streams via a Gap-based Weighting Approach (발생 간격 기반 가중치 부여 기법을 활용한 데이터 스트림에서 가중치 순차패턴 탐색)

  • Chang, Joong-Hyuk
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.55-75
    • /
    • 2010
  • Sequential pattern mining aims to discover interesting sequential patterns in a sequence database, and it is one of the essential data mining tasks widely used in various application fields such as Web access pattern analysis, customer purchase pattern analysis, and DNA sequence analysis. In general sequential pattern mining, only the generation order of data element in a sequence is considered, so that it can easily find simple sequential patterns, but has a limit to find more interesting sequential patterns being widely used in real world applications. One of the essential research topics to compensate the limit is a topic of weighted sequential pattern mining. In weighted sequential pattern mining, not only the generation order of data element but also its weight is considered to get more interesting sequential patterns. In recent, data has been increasingly taking the form of continuous data streams rather than finite stored data sets in various application fields, the database research community has begun focusing its attention on processing over data streams. The data stream is a massive unbounded sequence of data elements continuously generated at a rapid rate. In data stream processing, each data element should be examined at most once to analyze the data stream, and the memory usage for data stream analysis should be restricted finitely although new data elements are continuously generated in a data stream. Moreover, newly generated data elements should be processed as fast as possible to produce the up-to-date analysis result of a data stream, so that it can be instantly utilized upon request. To satisfy these requirements, data stream processing sacrifices the correctness of its analysis result by allowing some error. Considering the changes in the form of data generated in real world application fields, many researches have been actively performed to find various kinds of knowledge embedded in data streams. They mainly focus on efficient mining of frequent itemsets and sequential patterns over data streams, which have been proven to be useful in conventional data mining for a finite data set. In addition, mining algorithms have also been proposed to efficiently reflect the changes of data streams over time into their mining results. However, they have been targeting on finding naively interesting patterns such as frequent patterns and simple sequential patterns, which are found intuitively, taking no interest in mining novel interesting patterns that express the characteristics of target data streams better. Therefore, it can be a valuable research topic in the field of mining data streams to define novel interesting patterns and develop a mining method finding the novel patterns, which will be effectively used to analyze recent data streams. This paper proposes a gap-based weighting approach for a sequential pattern and amining method of weighted sequential patterns over sequence data streams via the weighting approach. A gap-based weight of a sequential pattern can be computed from the gaps of data elements in the sequential pattern without any pre-defined weight information. That is, in the approach, the gaps of data elements in each sequential pattern as well as their generation orders are used to get the weight of the sequential pattern, therefore it can help to get more interesting and useful sequential patterns. Recently most of computer application fields generate data as a form of data streams rather than a finite data set. Considering the change of data, the proposed method is mainly focus on sequence data streams.

Patent Analysis on 5G Technology Trends from the Perspective of Smart Factory (특허 분석을 통한 스마트공장 관점의 5G 기술개발 동향 연구)

  • Cho, Eunnuri;Chang, Tai-Woo
    • The Journal of Society for e-Business Studies
    • /
    • v.25 no.3
    • /
    • pp.95-108
    • /
    • 2020
  • The development of 5G technology, which is a next-generation communication technology capable of processing large amounts of data in real-time and solving delays, is drawing attention. Not only in the United States but also Korea, 5G is focused on supporting R&D as a national strategic technology. The strategy for the smart factory, one of the core services of the 5G, aims to increase the flexibility of manufacturing production lines. The existing wired communications devices can be replaced into wireless ones with the ultra-low-delay and ultra-high-speed characteristics of 5G. For the efficient development of 5G technology, it is necessary to keep abreast of the status and trend. In this study, based on the collected data of 1517 Korea patents and 1928 US patents, 5G technologies trend was analyzed and key technologies were identified by network analysis and topic modeling. We expect that it will be used for decision making for policy establishment and technology strategy of related industries to provide the trends of technology development related to the introduction of 5G technology to smart factories.

Eco-Friendly Light Emitting Diodes Based on Graphene Quantum Dots and III-V Colloidal Quantum Dots

  • Lee, Chang-Lyoul
    • Proceedings of the Korean Vacuum Society Conference
    • /
    • 2015.08a
    • /
    • pp.65-65
    • /
    • 2015
  • In this talk, I will introduce two topics. The first topic is the polymer light emitting diodes (PLEDs) using graphene oxide quantum dots as emissive center. More specifically, the energy transfer mechanism as well as the origin of white electroluminescence in the PLED were investigated. The second topic is the facile synthesis of eco-friendly III-V colloidal quantum dots and their application to light emitting diodes. Polymer (organic) light emitting diodes (PLEDs) using quantum dots (QDs) as emissive materials have received much attention as promising components for next-generation displays. Despite their outstanding properties, toxic and hazardous nature of QDs is a serious impediment to their use in future eco-friendly opto-electronic device applications. Owing to the desires to develop new types of nanomaterial without health and environmental effects but with strong opto-electrical properties similar to QDs, graphene quantum dots (GQDs) have attracted great interest as promising luminophores. However, the origin of electroluminescence (EL) from GQDs incorporated PLEDs is unclear. Herein, we synthesized graphene oxide quantum dots (GOQDs) using a modified hydrothermal deoxidization method and characterized the PLED performance using GOQDs blended poly(N-vinyl carbazole) (PVK) as emissive layer. Simple device structure was used to reveal the origin of EL by excluding the contribution of and contamination from other layers. The energy transfer and interaction between the PVK host and GOQDs guest were investigated using steady-state PL, time-correlated single photon counting (TCSPC) and density functional theory (DFT) calculations. Experiments revealed that white EL emission from the PLED originated from the hybridized GOQD-PVK complex emission with the contributions from the individual GOQDs and PVK emissions. (Sci Rep., 5, 11032, 2015). New III-V colloidal quantum dots (CQDs) were synthesized using the hot-injection method and the QD-light emitting diodes (QLEDs) using these CQDs as emissive layer were demonstrated for the first time. The band gaps of the III-V CQDs were varied by varying the metal fraction and by particle size control. The X-ray absorption fine structure (XAFS) results show that the crystal states of the III-V CQDs consist of multi-phase states; multi-peak photoluminescence (PL) resulted from these multi-phase states. Inverted structured QLED shows green EL emission and a maximum luminance of ~45 cd/m2. This result shows that III-V CQDs can be a good substitute for conventional cadmium-containing CQDs in various opto-electronic applications, e.g., eco-friendly displays. (Un-published results).

  • PDF

Mining Interesting Sequential Pattern with a Time-interval Constraint for Efficient Analyzing a Web-Click Stream (웹 클릭 스트림의 효율적 분석을 위한 시간 간격 제한을 활용한 관심 순차패턴 탐색)

  • Chang, Joong-Hyuk
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.16 no.2
    • /
    • pp.19-29
    • /
    • 2011
  • Due to the development of web technologies and the increasing use of smart devices such as smart phone, in recent various web services are widely used in many application fields. In this environment, the topic of supporting personalized and intelligent web services have been actively researched, and an analysis technique on a web-click stream generated from web usage logs is one of the essential techniques related to the topic. In this paper, for efficient analyzing a web-click stream of sequences, a sequential pattern mining technique is proposed, which satisfies the basic requirements for data stream processing and finds a refined mining result. For this purpose, a concept of interesting sequential patterns with a time-interval constraint is defined, which uses not on1y the order of items in a sequential pattern but also their generation times. In addition, A mining method to find the interesting sequential patterns efficiently over a data stream such as a web-click stream is proposed. The proposed method can be effectively used to various computing application fields such as E-commerce, bio-informatics, and USN environments, which generate data as a form of data streams.

Comparative Comprehension of Men Learning by the Principles of Complex System and the Book of Changes (복잡계의 원리와 주역의 사유방식이 주는 교육에의 시사점)

  • Park, Hye jeong;Do, Yeong ae
    • Korean Educational Research Journal
    • /
    • v.41 no.1
    • /
    • pp.59-79
    • /
    • 2020
  • Men learning has no fixed route. In other words, any route can be taken, which can also be seen in the structure of other "complex systems" discussed in modern society. It can also be examined through Yang's long-standing classic, The Book of Changes Men learning itself started from informal learning to become today's formal learning. As we look at the stages of human civilization's progress, we can quickly discover these stages of development. The issue of human beings has always been a topic of discussion, and these discussions are ongoing. Men learn through language and tools, technology and culture, and through philosophy, art, and religion to deal with their complex and diverse mental world. Through these various activities, learning is accomplished. This is not limited to the physical processes of one generation learning through inheriting knowledge; men's learning, a kind of mental process, has extended our life. This is why there is no other reason that men's minds and learning are always developing. This study is about how to learn in a complex and diversified modern society and to find out how to coexist with the principles of the "complex system" and The Book of Changes.

  • PDF

Eigen Palmprint Identification Algorithm using PCA(Principal Components Analysis) (주성분 분석법을 이용한 고유장문 인식 알고리즘)

  • Noh Jin-Soo;Rhee Kang-Hyeon
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.43 no.3 s.309
    • /
    • pp.82-89
    • /
    • 2006
  • Palmprint-based personal identification system, as a new member in the biometrics system family, has become an active research topic in recent years. Although lots of methods have been made, how to represent palmprint for effective classification is still an open problem and conducting researches. In this paper, the palmprint classification and recognition method based on PCA (Principal Components Analysis) using the dimension reduction of singular vector is proposed. And the 135dpi palmprint image which is obtained by the palmprint acquisition device is used for the effectual palmprint recognition system. The proposed system is consists of the palmprint acquisition device, DB generation algorithm and the palmprint recognition algorithm. The palmprint recognition step is limited 2 times. As a results GAR and FAR are 98.5% and 0.036%.