• Title/Summary/Keyword: 생성 데이터 증강

Search Result 138, Processing Time 0.02 seconds

A Study on Disease Prediction of Paralichthys Olivaceus using Deep Learning Technique (딥러닝 기술을 이용한 넙치의 질병 예측 연구)

  • Son, Hyun Seung;Lim, Han Kyu;Choi, Han Suk
    • Smart Media Journal
    • /
    • v.11 no.4
    • /
    • pp.62-68
    • /
    • 2022
  • To prevent the spread of disease in aquaculture, it is a need for a system to predict fish diseases while monitoring the water quality environment and the status of growing fish in real time. The existing research in predicting fish disease were image processing techniques. Recently, there have been more studies on disease prediction methods through deep learning techniques. This paper introduces the research results on how to predict diseases of Paralichthys Olivaceus with deep learning technology in aquaculture. The method enhances the performance of disease detection rates by including data augmentation and pre-processing in camera images collected from aquaculture. In this method, it is expected that early detection of disease fish will prevent fishery disasters such as mass closure of fish in aquaculture and reduce the damage of the spread of diseases to local aquaculture to prevent the decline in sales.

Deep-learning based SAR Ship Detection with Generative Data Augmentation (영상 생성적 데이터 증강을 이용한 딥러닝 기반 SAR 영상 선박 탐지)

  • Kwon, Hyeongjun;Jeong, Somi;Kim, SungTai;Lee, Jaeseok;Sohn, Kwanghoon
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.1
    • /
    • pp.1-9
    • /
    • 2022
  • Ship detection in synthetic aperture radar (SAR) images is an important application in marine monitoring for the military and civilian domains. Over the past decade, object detection has achieved significant progress with the development of convolutional neural networks (CNNs) and lot of labeled databases. However, due to difficulty in collecting and labeling SAR images, it is still a challenging task to solve SAR ship detection CNNs. To overcome the problem, some methods have employed conventional data augmentation techniques such as flipping, cropping, and affine transformation, but it is insufficient to achieve robust performance to handle a wide variety of types of ships. In this paper, we present a novel and effective approach for deep SAR ship detection, that exploits label-rich Electro-Optical (EO) images. The proposed method consists of two components: a data augmentation network and a ship detection network. First, we train the data augmentation network based on conditional generative adversarial network (cGAN), which aims to generate additional SAR images from EO images. Since it is trained using unpaired EO and SAR images, we impose the cycle-consistency loss to preserve the structural information while translating the characteristics of the images. After training the data augmentation network, we leverage the augmented dataset constituted with real and translated SAR images to train the ship detection network. The experimental results include qualitative evaluation of the translated SAR images and the comparison of detection performance of the networks, trained with non-augmented and augmented dataset, which demonstrates the effectiveness of the proposed framework.

Phase Segmentation of PVA Fiber-Reinforced Cementitious Composites Using U-net Deep Learning Approach (U-net 딥러닝 기법을 활용한 PVA 섬유 보강 시멘트 복합체의 섬유 분리)

  • Jeewoo Suh;Tong-Seok Han
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.36 no.5
    • /
    • pp.323-330
    • /
    • 2023
  • The development of an analysis model that reflects the microstructure characteristics of polyvinyl alcohol (PVA) fiber-reinforced cementitious composites, which have a highly complex microstructure, enables synergy between efficient material design and real experiments. PVA fiber orientations are an important factor that influences the mechanical behavior of PVA fiber-reinforced cementitious composites. Owing to the difficulty in distinguishing the gray level value obtained from micro-CT images of PVA fibers from adjacent phases, fiber segmentation is time-consuming work. In this study, a micro-CT test with a voxel size of 0.65 ㎛3 was performed to investigate the three-dimensional distribution of fibers. To segment the fibers and generate training data, histogram, morphology, and gradient-based phase-segmentation methods were used. A U-net model was proposed to segment fibers from micro-CT images of PVA fiber-reinforced cementitious composites. Data augmentation was applied to increase the accuracy of the training, using a total of 1024 images as training data. The performance of the model was evaluated using accuracy, precision, recall, and F1 score. The trained model achieved a high fiber segmentation performance and efficiency, and the approach can be applied to other specimens as well.

Development of Deep Recognition of Similarity in Show Garden Design Based on Deep Learning (딥러닝을 활용한 전시 정원 디자인 유사성 인지 모형 연구)

  • Cho, Woo-Yun;Kwon, Jin-Wook
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.52 no.2
    • /
    • pp.96-109
    • /
    • 2024
  • The purpose of this study is to propose a method for evaluating the similarity of Show gardens using Deep Learning models, specifically VGG-16 and ResNet50. A model for judging the similarity of show gardens based on VGG-16 and ResNet50 models was developed, and was referred to as DRG (Deep Recognition of similarity in show Garden design). An algorithm utilizing GAP and Pearson correlation coefficient was employed to construct the model, and the accuracy of similarity was analyzed by comparing the total number of similar images derived at 1st (Top1), 3rd (Top3), and 5th (Top5) ranks with the original images. The image data used for the DRG model consisted of a total of 278 works from the Le Festival International des Jardins de Chaumont-sur-Loire, 27 works from the Seoul International Garden Show, and 17 works from the Korea Garden Show. Image analysis was conducted using the DRG model for both the same group and different groups, resulting in the establishment of guidelines for assessing show garden similarity. First, overall image similarity analysis was best suited for applying data augmentation techniques based on the ResNet50 model. Second, for image analysis focusing on internal structure and outer form, it was effective to apply a certain size filter (16cm × 16cm) to generate images emphasizing form and then compare similarity using the VGG-16 model. It was suggested that an image size of 448 × 448 pixels and the original image in full color are the optimal settings. Based on these research findings, a quantitative method for assessing show gardens is proposed and it is expected to contribute to the continuous development of garden culture through interdisciplinary research moving forward.

Designing mobile personal assistant agent based on users' experience and their position information (위치정보 및 사용자 경험을 반영하는 모바일 PA에이전트의 설계)

  • Kang, Shin-Bong;Noh, Sang-Uk
    • Journal of Internet Computing and Services
    • /
    • v.12 no.1
    • /
    • pp.99-110
    • /
    • 2011
  • Mobile environments rapidly changing and digital convergence widely employed, mobile devices including smart phones have been playing a critical role that changes users' lifestyle in the areas of entertainments, businesses and information services. The various services using mobile devices are developing to meet the personal needs of users in the mobile environments. Especially, an LBS (Location-Based Service) is combined with other services and contents such as augmented reality, mobile SNS (Social Network Service), games, and searching, which can provide convenient and useful services to mobile users. In this paper, we design and implement the prototype of mobile personal assistant (PA) agents. Our personal assistant agent helps users do some tasks by hiding the complexity of difficult tasks, performing tasks on behalf of the users, and reflecting the preferences of users. To identify user's preferences and provide personalized services, clustering and classification algorithms of data mining are applied. The clusters of the log data using clustering algorithms are made by measuring the dissimilarity between two objects based on usage patterns. The classification algorithms produce user profiles within each cluster, which make it possible for PA agents to provide users with personalized services and contents. In the experiment, we measured the classification accuracy of user model clustered using clustering algorithms. It turned out that the classification accuracy using our method was increased by 17.42%, compared with that using other clustering algorithms.

Georeferencing of Indoor Omni-Directional Images Acquired by a Rotating Line Camera (회전식 라인 카메라로 획득한 실내 전방위 영상의 지오레퍼런싱)

  • Oh, So-Jung;Lee, Im-Pyeong
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.30 no.2
    • /
    • pp.211-221
    • /
    • 2012
  • To utilize omni-directional images acquired by a rotating line camera for indoor spatial information services, we should register precisely the images with respect to an indoor coordinate system. In this study, we thus develop a georeferencing method to estimate the exterior orientation parameters of an omni-directional image - the position and attitude of the camera at the acquisition time. First, we derive the collinearity equations for the omni-directional image by geometrically modeling the rotating line camera. We then estimate the exterior orientation parameters using the collinearity equations with indoor control points. The experimental results from the application to real data indicate that the exterior orientation parameters is estimated with the precision of 1.4 mm and $0.05^{\circ}$ for the position and attitude, respectively. The residuals are within 3 and 10 pixels in horizontal and vertical directions, respectively. Particularly, the residuals in the vertical direction retain systematic errors mainly due to the lens distortion, which should be eliminated through a camera calibration process. Using omni-directional images georeferenced precisely with the proposed method, we can generate high resolution indoor 3D models and sophisticated augmented reality services based on the models.

Annotation Method based on Face Area for Efficient Interactive Video Authoring (효과적인 인터랙티브 비디오 저작을 위한 얼굴영역 기반의 어노테이션 방법)

  • Yoon, Ui Nyoung;Ga, Myeong Hyeon;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.83-98
    • /
    • 2015
  • Many TV viewers use mainly portal sites in order to retrieve information related to broadcast while watching TV. However retrieving information that people wanted needs a lot of time to retrieve the information because current internet presents too much information which is not required. Consequentially, this process can't satisfy users who want to consume information immediately. Interactive video is being actively investigated to solve this problem. An interactive video provides clickable objects, areas or hotspots to interact with users. When users click object on the interactive video, they can see additional information, related to video, instantly. The following shows the three basic procedures to make an interactive video using interactive video authoring tool: (1) Create an augmented object; (2) Set an object's area and time to be displayed on the video; (3) Set an interactive action which is related to pages or hyperlink; However users who use existing authoring tools such as Popcorn Maker and Zentrick spend a lot of time in step (2). If users use wireWAX then they can save sufficient time to set object's location and time to be displayed because wireWAX uses vision based annotation method. But they need to wait for time to detect and track object. Therefore, it is required to reduce the process time in step (2) using benefits of manual annotation method and vision-based annotation method effectively. This paper proposes a novel annotation method allows annotator to easily annotate based on face area. For proposing new annotation method, this paper presents two steps: pre-processing step and annotation step. The pre-processing is necessary because system detects shots for users who want to find contents of video easily. Pre-processing step is as follow: 1) Extract shots using color histogram based shot boundary detection method from frames of video; 2) Make shot clusters using similarities of shots and aligns as shot sequences; and 3) Detect and track faces from all shots of shot sequence metadata and save into the shot sequence metadata with each shot. After pre-processing, user can annotates object as follow: 1) Annotator selects a shot sequence, and then selects keyframe of shot in the shot sequence; 2) Annotator annotates objects on the relative position of the actor's face on the selected keyframe. Then same objects will be annotated automatically until the end of shot sequence which has detected face area; and 3) User assigns additional information to the annotated object. In addition, this paper designs the feedback model in order to compensate the defects which are wrong aligned shots, wrong detected faces problem and inaccurate location problem might occur after object annotation. Furthermore, users can use interpolation method to interpolate position of objects which is deleted by feedback. After feedback user can save annotated object data to the interactive object metadata. Finally, this paper shows interactive video authoring system implemented for verifying performance of proposed annotation method which uses presented models. In the experiment presents analysis of object annotation time, and user evaluation. First, result of object annotation average time shows our proposed tool is 2 times faster than existing authoring tools for object annotation. Sometimes, annotation time of proposed tool took longer than existing authoring tools, because wrong shots are detected in the pre-processing. The usefulness and convenience of the system were measured through the user evaluation which was aimed at users who have experienced in interactive video authoring system. Recruited 19 experts evaluates of 11 questions which is out of CSUQ(Computer System Usability Questionnaire). CSUQ is designed by IBM for evaluating system. Through the user evaluation, showed that proposed tool is useful for authoring interactive video than about 10% of the other interactive video authoring systems.

A Survey of Yeosu Sado Dinosaur Tracksite and Utilization of Educational Materials using 3D Photogrammetry (3D 사진측량법을 이용한 여수 사도 공룡발자국 화석산지 조사 및 교육자료 활용방안)

  • Jo, Hyemin;Hong, Minsun;Son, Jongju;Lee, Hyun-Yeong;Park, Kyeong-Beom;Jung, Jongyun;Huh, Min
    • Journal of the Korean earth science society
    • /
    • v.42 no.6
    • /
    • pp.662-676
    • /
    • 2021
  • The Yeosu Sado dinosaur tracksite is well known for many dinosaur tracks and research on the gregarious behavior of dinosaurs. In addition, various geological and geographical heritage sites are distributed on Sado Island. However, educational field trips for students are very limited due to accessibility according to its geological location, time constraints due to tides, and continuous weathering and damage. Therefore, this study aims to generate 3D models and images of dinosaur tracks using the photogrammetric method, which has recently been used in various fields, and then discuss the possibility of using them as paleontological research and educational contents. As a result of checking the obtained 3D images and models, it was possible to confirm the existence of footprints that were not previously discovered or could not represent details by naked eyes or photos. Even previously discovered tracks could possibly present details using 3D images that could not be expressed by photos or interpretive drawings. In addition, the 3D model of dinosaur tracks can be preserved as semi-permanent data, enabling various forms of utilization and preservation. Here we apply 3D printing and mobile augmented reality content using photogrammetric 3D models for a virtual field trip, and these models acquired by photogrammetry can be used in various educational content fields that require 3D models.