• Title/Summary/Keyword: AI-based image generation

Search Result 37, Processing Time 0.019 seconds

Agricultural Applicability of AI based Image Generation (AI 기반 이미지 생성 기술의 농업 적용 가능성)

  • Seungri Yoon;Yeyeong Lee;Eunkyu Jung;Tae In Ahn
    • Journal of Bio-Environment Control
    • /
    • v.33 no.2
    • /
    • pp.120-128
    • /
    • 2024
  • Since ChatGPT was released in 2022, the generative artificial intelligence (AI) industry has seen massive growth and is expected to bring significant innovations to cognitive tasks. AI-based image generation, in particular, is leading major changes in the digital world. This study investigates the technical foundations of Midjourney, Stable Diffusion, and Firefly-three notable AI image generation tools-and compares their effectiveness by examining the images they produce. The results show that these AI tools can generate realistic images of tomatoes, strawberries, paprikas, and cucumbers, typical crops grown in greenhouse. Especially, Firefly stood out for its ability to produce very realistic images of greenhouse-grown crops. However, all tools struggled to fully capture the environmental context of greenhouses where these crops grow. The process of refining prompts and using reference images has proven effective in accurately generating images of strawberry fruits and their cultivation systems. In the case of generating cucumber images, the AI tools produced images very close to real ones, with no significant differences found in their evaluation scores. This study demonstrates how AI-based image generation technology can be applied in agriculture, suggesting a bright future for its use in this field.

A Study of an AI-Based Content Source Data Generation Model using Folk Paintings and Genre Paintings (민화와 풍속화를 이용한 AI 기반의 콘텐츠 원천 데이터 생성 모델의 연구)

  • Yang, Seokhwan;Lee, Young-Suk
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.5
    • /
    • pp.736-743
    • /
    • 2021
  • Due to COVID-19, the non-face-to-face content market is growing rapidly. However, most of the non-face-to-face content such as webtoons and web novels are produced based on the traditional culture of other countries, not Korean traditional culture. The biggest cause of this situation is the lack of reference materials for creating based on Korean traditional culture. Therefore, the need for materials on traditional Korean culture that can be used for content creation is emerging. In this paper, we propose a generation model of source data based on traditional folk paintings through the fusion of traditional Korean folk paintings and AI technology. The proposed model secures basic data based on folk tales, analyzes the style and characteristics of folk tales, and converts historical backgrounds and various stories related to folk tales into data. In addition, using the built data, various new stories are created based on AI technology. The proposed model is highly utilized in that it provides a foundation for new creation based on Korean traditional folk painting and AI technology.

Best Practice on Automatic Toon Image Creation from JSON File of Message Sequence Diagram via Natural Language based Requirement Specifications

  • Hyuntae Kim;Ji Hoon Kong;Hyun Seung Son;R. Young Chul Kim
    • International journal of advanced smart convergence
    • /
    • v.13 no.1
    • /
    • pp.99-107
    • /
    • 2024
  • In AI image generation tools, most general users must use an effective prompt to craft queries or statements to elicit the desired response (image, result) from the AI model. But we are software engineers who focus on software processes. At the process's early stage, we use informal and formal requirement specifications. At this time, we adapt the natural language approach into requirement engineering and toon engineering. Most Generative AI tools do not produce the same image in the same query. The reason is that the same data asset is not used for the same query. To solve this problem, we intend to use informal requirement engineering and linguistics to create a toon. Therefore, we propose a sequence diagram and image generation mechanism by analyzing and applying key objects and attributes as an informal natural language requirement analysis. Identify morpheme and semantic roles by analyzing natural language through linguistic methods. Based on the analysis results, a sequence diagram and an image are generated through the diagram. We expect consistent image generation using the same image element asset through the proposed mechanism.

Research on AI Painting Generation Technology Based on the [Stable Diffusion]

  • Chenghao Wang;Jeanhun Chung
    • International journal of advanced smart convergence
    • /
    • v.12 no.2
    • /
    • pp.90-95
    • /
    • 2023
  • With the rapid development of deep learning and artificial intelligence, generative models have achieved remarkable success in the field of image generation. By combining the stable diffusion method with Web UI technology, a novel solution is provided for the application of AI painting generation. The application prospects of this technology are very broad and can be applied to multiple fields, such as digital art, concept design, game development, and more. Furthermore, the platform based on Web UI facilitates user operations, making the technology more easily applicable to practical scenarios. This paper introduces the basic principles of Stable Diffusion Web UI technology. This technique utilizes the stability of diffusion processes to improve the output quality of generative models. By gradually introducing noise during the generation process, the model can generate smoother and more coherent images. Additionally, the analysis of different model types and applications within Stable Diffusion Web UI provides creators with a more comprehensive understanding, offering valuable insights for fields such as artistic creation and design.

A Case Study of Creative Art Based on AI Generation Technology

  • Qianqian Jiang;Jeanhun Chung
    • International journal of advanced smart convergence
    • /
    • v.12 no.2
    • /
    • pp.84-89
    • /
    • 2023
  • In recent years, with the breakthrough of Artificial Intelligence (AI) technology in deep learning algorithms such as Generative Adversarial Networks (GANs) and Variational Autoencoders (VAE), AI generation technology has rapidly expanded in various sub-sectors in the art field. 2022 as the explosive year of AI-generated art, especially in the creation of AI-generated art creative design, many excellent works have been born, which has improved the work efficiency of art design. This study analyzed the application design characteristics of AI generation technology in two sub fields of artistic creative design of AI painting and AI animation production , and compares the differences between traditional painting and AI painting in the field of painting. Through the research of this paper, the advantages and problems in the process of AI creative design are summarized. Although AI art designs are affected by technical limitations, there are still flaws in artworks and practical problems such as copyright and income, but it provides a strong technical guarantee in the expansion of subdivisions of artistic innovation and technology integration, and has extremely high research value.

Research on Generative AI for Korean Multi-Modal Montage App (한국형 멀티모달 몽타주 앱을 위한 생성형 AI 연구)

  • Lim, Jeounghyun;Cha, Kyung-Ae;Koh, Jaepil;Hong, Won-Kee
    • Journal of Service Research and Studies
    • /
    • v.14 no.1
    • /
    • pp.13-26
    • /
    • 2024
  • Multi-modal generation is the process of generating results based on a variety of information, such as text, images, and audio. With the rapid development of AI technology, there is a growing number of multi-modal based systems that synthesize different types of data to produce results. In this paper, we present an AI system that uses speech and text recognition to describe a person and generate a montage image. While the existing montage generation technology is based on the appearance of Westerners, the montage generation system developed in this paper learns a model based on Korean facial features. Therefore, it is possible to create more accurate and effective Korean montage images based on multi-modal voice and text specific to Korean. Since the developed montage generation app can be utilized as a draft montage, it can dramatically reduce the manual labor of existing montage production personnel. For this purpose, we utilized persona-based virtual person montage data provided by the AI-Hub of the National Information Society Agency. AI-Hub is an AI integration platform aimed at providing a one-stop service by building artificial intelligence learning data necessary for the development of AI technology and services. The image generation system was implemented using VQGAN, a deep learning model used to generate high-resolution images, and the KoDALLE model, a Korean-based image generation model. It can be confirmed that the learned AI model creates a montage image of a face that is very similar to what was described using voice and text. To verify the practicality of the developed montage generation app, 10 testers used it and more than 70% responded that they were satisfied. The montage generator can be used in various fields, such as criminal detection, to describe and image facial features.

Performance Comparison According to Image Generation Method in NIDS (Network Intrusion Detection System) using CNN

  • Sang Hyun, Kim
    • International journal of advanced smart convergence
    • /
    • v.12 no.2
    • /
    • pp.67-75
    • /
    • 2023
  • Recently, many studies have been conducted on ways to utilize AI technology in NIDS (Network Intrusion Detection System). In particular, CNN-based NIDS generally shows excellent performance. CNN is basically a method of using correlation between pixels existing in an image. Therefore, the method of generating an image is very important in CNN. In this paper, the performance comparison of CNN-based NIDS according to the image generation method was performed. The image generation methods used in the experiment are a direct conversion method and a one-hot encoding based method. As a result of the experiment, the performance of NIDS was different depending on the image generation method. In particular, it was confirmed that the method combining the direct conversion method and the one-hot encoding based method proposed in this paper showed the best performance.

A Comparative Analysis Between <Leonardo.Ai> and <Meshy> as AI Texture Generation Tools

  • Pingjian Jie;Xinyi Shan;Jeanhun Chung
    • International Journal of Advanced Culture Technology
    • /
    • v.11 no.4
    • /
    • pp.333-339
    • /
    • 2023
  • In three-dimensional(3D) modeling, texturing plays a crucial role as a visual element, imparting detail and realism to models. In contrast to traditional texturing methods, the current trend involves utilizing AI tools such as Leonardo.Ai and Meshy to create textures for 3D models in a more efficient and precise manner. This paper focuses on 3D texturing, conducting a comprehensive comparative study of AI tools, specifically Leonardo.Ai and Meshy. By delving into the performance, functional differences, and respective application scopes of these two tools in the generation of 3D textures, we highlight potential applications and development trends within the realm of 3D texturing. The efficient use of AI tools in texture creation also has the potential to drive innovation and enhancement in the field of 3D modeling. In conclusion, this research aims to provide a comprehensive perspective for researchers, practitioners, and enthusiasts in related fields, fostering further innovation and development in this domain.

A Feasibility Study on RUNWAY GEN-2 for Generating Realistic Style Images

  • Yifan Cui;Xinyi Shan;Jeanhun Chung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.1
    • /
    • pp.99-105
    • /
    • 2024
  • Runway released an updated version, Gen-2, in March 2023, which introduced new features that are different from Gen-1: it can convert text and images into videos, or convert text and images together into video images based on text instructions. This update will be officially open to the public in June 2023, so more people can enjoy and use their creativity. With this new feature, users can easily transform text and images into impressive video creations. However, as with all new technologies, comes the instability of AI, which also affects the results generated by Runway. This article verifies the feasibility of using Runway to generate the desired video from several aspects through personal practice. In practice, I discovered Runway generation problems and propose improvement methods to find ways to improve the accuracy of Runway generation. And found that although the instability of AI is a factor that needs attention, through careful adjustment and testing, users can still make full use of this feature and create stunning video works. This update marks the beginning of a more innovative and diverse future for the digital creative field.