• Title/Summary/Keyword: Character Generation System

Search Result 104, Processing Time 0.019 seconds

The way to make training data for deep learning model to recognize keywords in product catalog image at E-commerce (온라인 쇼핑몰에서 상품 설명 이미지 내의 키워드 인식을 위한 딥러닝 훈련 데이터 자동 생성 방안)

  • Kim, Kitae;Oh, Wonseok;Lim, Geunwon;Cha, Eunwoo;Shin, Minyoung;Kim, Jongwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.1-23
    • /
    • 2018
  • From the 21st century, various high-quality services have come up with the growth of the internet or 'Information and Communication Technologies'. Especially, the scale of E-commerce industry in which Amazon and E-bay are standing out is exploding in a large way. As E-commerce grows, Customers could get what they want to buy easily while comparing various products because more products have been registered at online shopping malls. However, a problem has arisen with the growth of E-commerce. As too many products have been registered, it has become difficult for customers to search what they really need in the flood of products. When customers search for desired products with a generalized keyword, too many products have come out as a result. On the contrary, few products have been searched if customers type in details of products because concrete product-attributes have been registered rarely. In this situation, recognizing texts in images automatically with a machine can be a solution. Because bulk of product details are written in catalogs as image format, most of product information are not searched with text inputs in the current text-based searching system. It means if information in images can be converted to text format, customers can search products with product-details, which make them shop more conveniently. There are various existing OCR(Optical Character Recognition) programs which can recognize texts in images. But existing OCR programs are hard to be applied to catalog because they have problems in recognizing texts in certain circumstances, like texts are not big enough or fonts are not consistent. Therefore, this research suggests the way to recognize keywords in catalog with the Deep Learning algorithm which is state of the art in image-recognition area from 2010s. Single Shot Multibox Detector(SSD), which is a credited model for object-detection performance, can be used with structures re-designed to take into account the difference of text from object. But there is an issue that SSD model needs a lot of labeled-train data to be trained, because of the characteristic of deep learning algorithms, that it should be trained by supervised-learning. To collect data, we can try labelling location and classification information to texts in catalog manually. But if data are collected manually, many problems would come up. Some keywords would be missed because human can make mistakes while labelling train data. And it becomes too time-consuming to collect train data considering the scale of data needed or costly if a lot of workers are hired to shorten the time. Furthermore, if some specific keywords are needed to be trained, searching images that have the words would be difficult, as well. To solve the data issue, this research developed a program which create train data automatically. This program can make images which have various keywords and pictures like catalog and save location-information of keywords at the same time. With this program, not only data can be collected efficiently, but also the performance of SSD model becomes better. The SSD model recorded 81.99% of recognition rate with 20,000 data created by the program. Moreover, this research had an efficiency test of SSD model according to data differences to analyze what feature of data exert influence upon the performance of recognizing texts in images. As a result, it is figured out that the number of labeled keywords, the addition of overlapped keyword label, the existence of keywords that is not labeled, the spaces among keywords and the differences of background images are related to the performance of SSD model. This test can lead performance improvement of SSD model or other text-recognizing machine based on deep learning algorithm with high-quality data. SSD model which is re-designed to recognize texts in images and the program developed for creating train data are expected to contribute to improvement of searching system in E-commerce. Suppliers can put less time to register keywords for products and customers can search products with product-details which is written on the catalog.

A Study on Practical Curriculum Development of the Education for Love based on the Understanding of Psychoanalytic 'Desire of Subject' (정신분석학적 '욕망의 주체' 이해에 기초한 사랑의 교육 교육과정 개발)

  • Kim, Sun Ah
    • Journal of Christian Education in Korea
    • /
    • v.68
    • /
    • pp.77-112
    • /
    • 2021
  • This study is based on the research of the first year, which is the National Research Foundation of Korea's R&D subject for middle-grade researchers. In this study, the practical curriculum development of the education for love - an according to the psychoanalytic perspectives of F. Dolto(1908-1988) - is suggested as follows. The first is 'the reconstruction of the directions of curriculum and its specific aims in accordance with such directions.' The reconstruction of the directions of curriculum aims at leading our future generation to live as a subject of desire through the mutual-communication of love. The second is 'the reconstruction of the tasks of curriculum and its specific contents in accordance with such tasks.' The reconstruction of the tasks of curriculum pursuit to help our future generation through the converting the education for love into the paradigm of desire of Agape to live as a subject of desire forming a whole personality and practicing the desire of Agape in daily life. as a source of desire. According to these aims, the reconstruction of directions of curriculum are presented as following: firstly, 'curriculum for the mutual-communication between subjects of love' and secondly, 'curriculum for the subject of desire'. The reconstruction of tasks of curriculum are like these: firstly, 'converting the education for love into the paradigm of desire of Agape', and secondly, 'forming a whole personality through the education for love'. Thus, with respect to two specific aims in accordance with the reconstruction of directions are suggested like these: Firstly, 'constructing a subject as a speaking existence' and secondly, 'realizing the subject as the autonomous source of desire'. In the two specific contents in accordance with the reconstruction of tasks are presented as following: Firstly, 'realizing the truth of the desire of Agape'.' Secondly, 'practicing the desire of Agape in daily life.' The third is 'the reconstruction of curriculum by life cycle' are suggested. They include the fetal life, infants and preschool children life, and childhood life. In further study, it is required to contain adolescent period. It will be useful to help them to recover their self-esteem with the experience of true love, especially, out-of-school young generation overcome negative perspectives and prejudice in the society, and challenges to their dreams and future through proper utilization of the study outcome. The outcome of this study, which presented practical curriculum development of the education for love based on the understanding of psychoanalytic 'desire of subject' can be used as basic teaching materials for our future generations. Furthermore, the results can be used as a resource for educating ministers and lay leaders in the religious world to build capabilities to heal their inner side as well as the society that is tainted with various forms of conflict. These include general conflicts, anger, pleasure and addiction, depression and suicide, violence and murder, etc. The study outcome can contribute to the prevention of antisocial incidents against humanity that have recently been occurring in our free-semester system implemented in all middle schools across the country to be operated effectively. For example, it is possible to provide the study results as lecture and teaching materials for 'character camp' (self-examination and self-esteem improvement training) and 'family healing camp' (solution of a communication gap between family members and love communication training), which help students participate in field trip activities and career exploration activities voluntarily, independently, and creatively. Ultimately, it can visibly present the convergent research performance by providing the study outcome as preliminary data for the development of lecture videos and materials including infant care and preschool education, parental education, family consultation education, and holistic healing education. Support from the religious world, including the central government and local governments, are urgently required in order for such educational possibilities to be fulfilled both in the society and the fields of church education and to be actively linked to follow-up studies.

Deep Learning-based Professional Image Interpretation Using Expertise Transplant (전문성 이식을 통한 딥러닝 기반 전문 이미지 해석 방법론)

  • Kim, Taejin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.79-104
    • /
    • 2020
  • Recently, as deep learning has attracted attention, the use of deep learning is being considered as a method for solving problems in various fields. In particular, deep learning is known to have excellent performance when applied to applying unstructured data such as text, sound and images, and many studies have proven its effectiveness. Owing to the remarkable development of text and image deep learning technology, interests in image captioning technology and its application is rapidly increasing. Image captioning is a technique that automatically generates relevant captions for a given image by handling both image comprehension and text generation simultaneously. In spite of the high entry barrier of image captioning that analysts should be able to process both image and text data, image captioning has established itself as one of the key fields in the A.I. research owing to its various applicability. In addition, many researches have been conducted to improve the performance of image captioning in various aspects. Recent researches attempt to create advanced captions that can not only describe an image accurately, but also convey the information contained in the image more sophisticatedly. Despite many recent efforts to improve the performance of image captioning, it is difficult to find any researches to interpret images from the perspective of domain experts in each field not from the perspective of the general public. Even for the same image, the part of interests may differ according to the professional field of the person who has encountered the image. Moreover, the way of interpreting and expressing the image also differs according to the level of expertise. The public tends to recognize the image from a holistic and general perspective, that is, from the perspective of identifying the image's constituent objects and their relationships. On the contrary, the domain experts tend to recognize the image by focusing on some specific elements necessary to interpret the given image based on their expertise. It implies that meaningful parts of an image are mutually different depending on viewers' perspective even for the same image. So, image captioning needs to implement this phenomenon. Therefore, in this study, we propose a method to generate captions specialized in each domain for the image by utilizing the expertise of experts in the corresponding domain. Specifically, after performing pre-training on a large amount of general data, the expertise in the field is transplanted through transfer-learning with a small amount of expertise data. However, simple adaption of transfer learning using expertise data may invoke another type of problems. Simultaneous learning with captions of various characteristics may invoke so-called 'inter-observation interference' problem, which make it difficult to perform pure learning of each characteristic point of view. For learning with vast amount of data, most of this interference is self-purified and has little impact on learning results. On the contrary, in the case of fine-tuning where learning is performed on a small amount of data, the impact of such interference on learning can be relatively large. To solve this problem, therefore, we propose a novel 'Character-Independent Transfer-learning' that performs transfer learning independently for each character. In order to confirm the feasibility of the proposed methodology, we performed experiments utilizing the results of pre-training on MSCOCO dataset which is comprised of 120,000 images and about 600,000 general captions. Additionally, according to the advice of an art therapist, about 300 pairs of 'image / expertise captions' were created, and the data was used for the experiments of expertise transplantation. As a result of the experiment, it was confirmed that the caption generated according to the proposed methodology generates captions from the perspective of implanted expertise whereas the caption generated through learning on general data contains a number of contents irrelevant to expertise interpretation. In this paper, we propose a novel approach of specialized image interpretation. To achieve this goal, we present a method to use transfer learning and generate captions specialized in the specific domain. In the future, by applying the proposed methodology to expertise transplant in various fields, we expected that many researches will be actively conducted to solve the problem of lack of expertise data and to improve performance of image captioning.

Discussions about Expanded Fests of Cartoons and Multimedia Comics as Visual Culture: With a Focus on New Technologies (비주얼 컬처로서 만화영상의 확장된 장(場, fest)에 대한 논의: 뉴 테크놀로지를 중심으로)

  • Lee, Hwa-Ja;Kim, Se-Jong
    • Cartoon and Animation Studies
    • /
    • s.28
    • /
    • pp.1-25
    • /
    • 2012
  • The rapid digitalization across all aspects of society since 1990 led to the digitalization of cartoons. As the medium of cartoons moved from paper to the web, a powerful visual culture emerged. An encounter between cartoons and multimedia technologies has helped cartoons evolve into a video culture. Today cartoons are no longer literate culture. It is critical to pay attention to cartoons as an "expanded fest" and as visual and video culture with much broader significance. In this paper, the investigator set out to diagnose the current position of cartoons changing in the rapidly changing digital age and talk about future directions that they should pursue. Thus she discussed cases of changes from 1990 when colleges began to provide specialized education for cartoons and animation to the present day when cartoon and Multimedia Comics fests exist in addition to the digitalization of cartoons. The encounter between new technologies and cartoons broke down the conventional forms of cartoons. The massive appearance of artists that made active use of new technologies in their works, in particular, has facilitated changes to the content and forms of cartoons and the expansion of character uses. The development of high technologies extends influence to the roles of appreciators beyond the artists' works. Today readers voice their opinions about works actively, build a fan base, promote the works and artists they favor, and help them rise to stardom. As artist groups of various genres were formed, the possibilities of new stories and texts and the appearance of diverse styles and world views have expanded the essence of cartoon texts and the overall cartoon system of cartoon culture, industry, education, institution, and technology. It is expected that cartoons and Multimedia Comics will continue to make a contribution as a messenger to reflect the next generation of culture, mediate it, and communicate with it. Today there is no longer a distinction between print and video cartoons. Cartoons will expand in every field through a wide range of forms and styles, given the current situations involving installation concept cartoons, blockbuster digital videos, fancy items, and characters at theme parks based on a narrative. It is therefore necessary to diversify cartoon and Multimedia Comics education in diverse ways. Today educators are faced with a task to bring up future generations of talents who are capable of leading the culture of overall senses based on literate and video culture by incorporating humanities, social studies, and new technology education into their creative artistic abilities.