• Title/Summary/Keyword: module learning

Search Result 549, Processing Time 0.025 seconds

A study on the development of surveillance system for multiple drones in school drone education sites (학내 드론 교육현장의 다중드론 감시시스템 개발에 관한 연구)

  • Jin-Taek Lim;Sung-goo Yoo
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.1
    • /
    • pp.697-702
    • /
    • 2023
  • Recently, with the introduction of drones, a core technology of the 4th industrial revolution, various convergence education using drones is being conducted in school education sites. In particular, drone theory and practice education is being conducted in connection with free semester classes and career exploration. The drone convergence education program has higher learner satisfaction than simple demonstration and practice education, and the learning effect is high due to direct practical experience. However, since practical education is being conducted for a large number of learners, it is impossible to restrict and control the flight of a large number of drones in a limited place. In this paper, we propose a monitoring system that allows the instructor to monitor multiple drones in real time and learners to recognize collisions between drones in advance when multiple drones are operated, focusing on education operated in schools. The communication module used in the experiment was equipped with GPS in Murata LoRa, and the server and client were configured to enable monitoring based on the location data received in real time. The performance of the proposed system was evaluated in an open space, and it was confirmed that the communication signal was good up to a distance of about 120m. In other words, it was confirmed that 25 educational drones can be controlled within a range of 240m and the instructor can monitor them.

A Study on the Improvement of the Cadastral Field Process Evaluation-type Qualification System (지적분야 과정평가형 자격제도 개선에 관한 연구)

  • Seo, Yong-Su
    • Journal of Cadastre & Land InformatiX
    • /
    • v.53 no.1
    • /
    • pp.5-20
    • /
    • 2023
  • The government is operating a course evaluation-type qualification system for the purpose of overcoming the limitations of the test-type qualification system and cultivating talents with practical skills applicable to related industries. However, there is a part that does not match the purpose of the course evaluation qualification system by educating competency units in the field of surveying rather than the field of cadastral field in the education and training course of the course evaluation qualification. Therefore, this study closely analyzed the course evaluation qualification system in the cadastral field and suggested improvement plans accordingly. First, NCS in the cadastral field has not been supplemented or developed since its development in 2014, and there are many parts that do not conform to the current laws, so supplementary development is necessary. Second, It was proposed to improve the optional competency unit in the field of surveying to the NCS competency unit in the cadastral field. In addition, if the 12 competency units of the current NCS in the cadastral field are designated as essential competency units and education and training are operated, it will be possible to foster working-level talents with sufficient competency at the cadastral engineer level. However, for practical system operation and application, supplementation and development of NCS in the cadastral field, which is used for education and training of the course evaluation qualification system, must be preceded.

Development of Intelligent OCR Technology to Utilize Document Image Data (문서 이미지 데이터 활용을 위한 지능형 OCR 기술 개발)

  • Kim, Sangjun;Yu, Donghui;Hwang, Soyoung;Kim, Minho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.212-215
    • /
    • 2022
  • In the era of so-called digital transformation today, the need for the construction and utilization of big data in various fields has increased. Today, a lot of data is produced and stored in a digital device and media-friendly manner, but the production and storage of data for a long time in the past has been dominated by print books. Therefore, the need for Optical Character Recognition (OCR) technology to utilize the vast amount of print books accumulated for a long time as big data was also required in line with the need for big data. In this study, a system for digitizing the structure and content of a document object inside a scanned book image is proposed. The proposal system largely consists of the following three steps. 1) Recognition of area information by document objects (table, equation, picture, text body) in scanned book image. 2) OCR processing for each area of the text body-table-formula module according to recognized document object areas. 3) The processed document informations gather up and returned to the JSON format. The model proposed in this study uses an open-source project that additional learning and improvement. Intelligent OCR proposed as a system in this study showed commercial OCR software-level performance in processing four types of document objects(table, equation, image, text body).

  • PDF

A Study on Market Size Estimation Method by Product Group Using Word2Vec Algorithm (Word2Vec을 활용한 제품군별 시장규모 추정 방법에 관한 연구)

  • Jung, Ye Lim;Kim, Ji Hui;Yoo, Hyoung Sun
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.1-21
    • /
    • 2020
  • With the rapid development of artificial intelligence technology, various techniques have been developed to extract meaningful information from unstructured text data which constitutes a large portion of big data. Over the past decades, text mining technologies have been utilized in various industries for practical applications. In the field of business intelligence, it has been employed to discover new market and/or technology opportunities and support rational decision making of business participants. The market information such as market size, market growth rate, and market share is essential for setting companies' business strategies. There has been a continuous demand in various fields for specific product level-market information. However, the information has been generally provided at industry level or broad categories based on classification standards, making it difficult to obtain specific and proper information. In this regard, we propose a new methodology that can estimate the market sizes of product groups at more detailed levels than that of previously offered. We applied Word2Vec algorithm, a neural network based semantic word embedding model, to enable automatic market size estimation from individual companies' product information in a bottom-up manner. The overall process is as follows: First, the data related to product information is collected, refined, and restructured into suitable form for applying Word2Vec model. Next, the preprocessed data is embedded into vector space by Word2Vec and then the product groups are derived by extracting similar products names based on cosine similarity calculation. Finally, the sales data on the extracted products is summated to estimate the market size of the product groups. As an experimental data, text data of product names from Statistics Korea's microdata (345,103 cases) were mapped in multidimensional vector space by Word2Vec training. We performed parameters optimization for training and then applied vector dimension of 300 and window size of 15 as optimized parameters for further experiments. We employed index words of Korean Standard Industry Classification (KSIC) as a product name dataset to more efficiently cluster product groups. The product names which are similar to KSIC indexes were extracted based on cosine similarity. The market size of extracted products as one product category was calculated from individual companies' sales data. The market sizes of 11,654 specific product lines were automatically estimated by the proposed model. For the performance verification, the results were compared with actual market size of some items. The Pearson's correlation coefficient was 0.513. Our approach has several advantages differing from the previous studies. First, text mining and machine learning techniques were applied for the first time on market size estimation, overcoming the limitations of traditional sampling based- or multiple assumption required-methods. In addition, the level of market category can be easily and efficiently adjusted according to the purpose of information use by changing cosine similarity threshold. Furthermore, it has a high potential of practical applications since it can resolve unmet needs for detailed market size information in public and private sectors. Specifically, it can be utilized in technology evaluation and technology commercialization support program conducted by governmental institutions, as well as business strategies consulting and market analysis report publishing by private firms. The limitation of our study is that the presented model needs to be improved in terms of accuracy and reliability. The semantic-based word embedding module can be advanced by giving a proper order in the preprocessed dataset or by combining another algorithm such as Jaccard similarity with Word2Vec. Also, the methods of product group clustering can be changed to other types of unsupervised machine learning algorithm. Our group is currently working on subsequent studies and we expect that it can further improve the performance of the conceptually proposed basic model in this study.

A Study on Differences of Contents and Tones of Arguments among Newspapers Using Text Mining Analysis (텍스트 마이닝을 활용한 신문사에 따른 내용 및 논조 차이점 분석)

  • Kam, Miah;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.53-77
    • /
    • 2012
  • This study analyses the difference of contents and tones of arguments among three Korean major newspapers, the Kyunghyang Shinmoon, the HanKyoreh, and the Dong-A Ilbo. It is commonly accepted that newspapers in Korea explicitly deliver their own tone of arguments when they talk about some sensitive issues and topics. It could be controversial if readers of newspapers read the news without being aware of the type of tones of arguments because the contents and the tones of arguments can affect readers easily. Thus it is very desirable to have a new tool that can inform the readers of what tone of argument a newspaper has. This study presents the results of clustering and classification techniques as part of text mining analysis. We focus on six main subjects such as Culture, Politics, International, Editorial-opinion, Eco-business and National issues in newspapers, and attempt to identify differences and similarities among the newspapers. The basic unit of text mining analysis is a paragraph of news articles. This study uses a keyword-network analysis tool and visualizes relationships among keywords to make it easier to see the differences. Newspaper articles were gathered from KINDS, the Korean integrated news database system. KINDS preserves news articles of the Kyunghyang Shinmun, the HanKyoreh and the Dong-A Ilbo and these are open to the public. This study used these three Korean major newspapers from KINDS. About 3,030 articles from 2008 to 2012 were used. International, national issues and politics sections were gathered with some specific issues. The International section was collected with the keyword of 'Nuclear weapon of North Korea.' The National issues section was collected with the keyword of '4-major-river.' The Politics section was collected with the keyword of 'Tonghap-Jinbo Dang.' All of the articles from April 2012 to May 2012 of Eco-business, Culture and Editorial-opinion sections were also collected. All of the collected data were handled and edited into paragraphs. We got rid of stop-words using the Lucene Korean Module. We calculated keyword co-occurrence counts from the paired co-occurrence list of keywords in a paragraph. We made a co-occurrence matrix from the list. Once the co-occurrence matrix was built, we used the Cosine coefficient matrix as input for PFNet(Pathfinder Network). In order to analyze these three newspapers and find out the significant keywords in each paper, we analyzed the list of 10 highest frequency keywords and keyword-networks of 20 highest ranking frequency keywords to closely examine the relationships and show the detailed network map among keywords. We used NodeXL software to visualize the PFNet. After drawing all the networks, we compared the results with the classification results. Classification was firstly handled to identify how the tone of argument of a newspaper is different from others. Then, to analyze tones of arguments, all the paragraphs were divided into two types of tones, Positive tone and Negative tone. To identify and classify all of the tones of paragraphs and articles we had collected, supervised learning technique was used. The Na$\ddot{i}$ve Bayesian classifier algorithm provided in the MALLET package was used to classify all the paragraphs in articles. After classification, Precision, Recall and F-value were used to evaluate the results of classification. Based on the results of this study, three subjects such as Culture, Eco-business and Politics showed some differences in contents and tones of arguments among these three newspapers. In addition, for the National issues, tones of arguments on 4-major-rivers project were different from each other. It seems three newspapers have their own specific tone of argument in those sections. And keyword-networks showed different shapes with each other in the same period in the same section. It means that frequently appeared keywords in articles are different and their contents are comprised with different keywords. And the Positive-Negative classification showed the possibility of classifying newspapers' tones of arguments compared to others. These results indicate that the approach in this study is promising to be extended as a new tool to identify the different tones of arguments of newspapers.

Development of NCS Based Vocational Curriculum Model for the Practical and Creative Human Respirces (실전 창의형 인재 양성을 위한 NCS 기반 직업교육과정의 모형 개발)

  • Kim, Dong-Yeon;Kim, Jinsoo
    • 대한공업교육학회지
    • /
    • v.39 no.2
    • /
    • pp.101-121
    • /
    • 2014
  • The study aims to develop the NCS based vocational curriculum model for the practical and creative human resources. For effectiveness of the study, the study consists of literature studies of both domestic and international, contents analysis, case study, expert(9samples) consultation and review, and in-depth-interview of the three advisory members. The validity of the developed model is analyzed through mean, standard deviation and contents validity ratio(CVR). The main results of the model development in our study are as follow. First, our NCS based vocational curriculum model for the practical and creative human resources is developed with the analyses of NCS development manuals, training standard utilization and training curriculum organization manuals, NCS learning module development manual and case studies, NCS research report, NCS based curriculum pilot development resources directed toward the high schools and vocational school as well as the domestic and international literature study on career training model like NCS. Second, based on the findings of our analysis in combination with the findings from the consultations with the expert and advisory committee, total 19 sub-factors of each step and domain are extracted. The sub-factors of domain in step 1 are the competency unit, definition of competency unit, competency unit element, performance criteria, range of variable, guide of assessment, key competency; in step 2, they are subject title, subject objectives, chapter title, chapter objectives, pedagogical methods, assessment methods and basic job competence; and in step 2, they are NCS based subject matrix table, NCS based subject profile, NCS based job training curriculum table, NCS based subjects organization flowchart, NCS based job training operation plan. Third, the final model including step 3 NCS based subject profile are developed in association with the linked organizational sub-factors of step 1 and step 2. Forth, the validity tests for the final model by the step and domain yield the mean 4.67, CVR value 1.00, indicating the superior validity. Also, the means of each sub-factors are all over 4.33 with the CVR value 1.00, indicating the high validity as well. The means of the associated organizations within the model are also over 4.33 with the CVR value of 1.00. Standard deviations are all .50 or lower which are small. Fifth, based on the validity test results and the in-depth-interview of the expert and advisory committee, the model is adjusted complemented to establish final model of the NCS based vocational curriculum for the practical and creative human resources.

A Methodology for Automatic Multi-Categorization of Single-Categorized Documents (단일 카테고리 문서의 다중 카테고리 자동확장 방법론)

  • Hong, Jin-Sung;Kim, Namgyu;Lee, Sangwon
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.3
    • /
    • pp.77-92
    • /
    • 2014
  • Recently, numerous documents including unstructured data and text have been created due to the rapid increase in the usage of social media and the Internet. Each document is usually provided with a specific category for the convenience of the users. In the past, the categorization was performed manually. However, in the case of manual categorization, not only can the accuracy of the categorization be not guaranteed but the categorization also requires a large amount of time and huge costs. Many studies have been conducted towards the automatic creation of categories to solve the limitations of manual categorization. Unfortunately, most of these methods cannot be applied to categorizing complex documents with multiple topics because the methods work by assuming that one document can be categorized into one category only. In order to overcome this limitation, some studies have attempted to categorize each document into multiple categories. However, they are also limited in that their learning process involves training using a multi-categorized document set. These methods therefore cannot be applied to multi-categorization of most documents unless multi-categorized training sets are provided. To overcome the limitation of the requirement of a multi-categorized training set by traditional multi-categorization algorithms, we propose a new methodology that can extend a category of a single-categorized document to multiple categorizes by analyzing relationships among categories, topics, and documents. First, we attempt to find the relationship between documents and topics by using the result of topic analysis for single-categorized documents. Second, we construct a correspondence table between topics and categories by investigating the relationship between them. Finally, we calculate the matching scores for each document to multiple categories. The results imply that a document can be classified into a certain category if and only if the matching score is higher than the predefined threshold. For example, we can classify a certain document into three categories that have larger matching scores than the predefined threshold. The main contribution of our study is that our methodology can improve the applicability of traditional multi-category classifiers by generating multi-categorized documents from single-categorized documents. Additionally, we propose a module for verifying the accuracy of the proposed methodology. For performance evaluation, we performed intensive experiments with news articles. News articles are clearly categorized based on the theme, whereas the use of vulgar language and slang is smaller than other usual text document. We collected news articles from July 2012 to June 2013. The articles exhibit large variations in terms of the number of types of categories. This is because readers have different levels of interest in each category. Additionally, the result is also attributed to the differences in the frequency of the events in each category. In order to minimize the distortion of the result from the number of articles in different categories, we extracted 3,000 articles equally from each of the eight categories. Therefore, the total number of articles used in our experiments was 24,000. The eight categories were "IT Science," "Economy," "Society," "Life and Culture," "World," "Sports," "Entertainment," and "Politics." By using the news articles that we collected, we calculated the document/category correspondence scores by utilizing topic/category and document/topics correspondence scores. The document/category correspondence score can be said to indicate the degree of correspondence of each document to a certain category. As a result, we could present two additional categories for each of the 23,089 documents. Precision, recall, and F-score were revealed to be 0.605, 0.629, and 0.617 respectively when only the top 1 predicted category was evaluated, whereas they were revealed to be 0.838, 0.290, and 0.431 when the top 1 - 3 predicted categories were considered. It was very interesting to find a large variation between the scores of the eight categories on precision, recall, and F-score.

Development and Complementation of Evaluation Area and Content Elements in Electrical, Electronics and Communications Subject (중등교사 임용후보자선정경쟁시험 표시과목인 전기·전자·통신의 평가영역 및 내용요소 개발·보완 연구)

  • Song, Youngjik;Kang, Yoonkook;Cho, Hanwook;Gim, Seongdeuk;Lim, Seunggak;Lee, Hyuksoo
    • 대한공업교육학회지
    • /
    • v.44 no.1
    • /
    • pp.52-71
    • /
    • 2019
  • The quality of school education is a key element for national education development. An important factor that determines the quality of school education is qualities of teachers who are in responsible for school education in the field. Therefore, it is necessary to hire competent teachers in the teacher appointment exam for the secondary school. This necessity is evident especially for vocational high schools and Meister high schools with the introduction of 2015-revised curriculum based on NCS that separates each three subjects, "Electrical, Electronics Communication" resulting in the change of question mechanism, which requires new designing of assessment and content area. So, this study analyzes curriculum in college of education for "Electrical", "Electronics", "Communication", 2015-revised curriculum based on NCS and the development of standards for teacher qualifications and assessment area and evaluation of teaching ability in the subjects of the teacher appointment exam, "Electrical, Electronics Communication" Engineering" in 2009. The assessment area and content elements of "Electrical", "Electronics", "Communication are extracted from the analyzed results and they are verified by experts' consultation and presented as follows; First, the assessment area and content elements of the "Electrical" subject were designed to evaluate the NCS - based 2015 revised curriculum by presenting the NCS learning module to the evaluation area and content element in the basic subject "Electrical and Electronics Practice". Second, the section of "Electronics" presented the assessment area and content elements applying the Electronic Circuit, basic subject of the NCS and it also added "Electromagnetics", which is the basic part of Electronics in the Application of Electromagnetic waves that could be applied to the assessment. Third, the assessment area and content elements of "Communication" consist of the communication-related practice that is based on "Electrical" and "Electronic", considering the characteristics of "Communication Engineering". In particular, "Electrical and Electronics practice" which adds network construction practice and communication-related practice makes it to be able to evaluate the communication-related practical education.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.