• 제목/요약/키워드: Generative Adversarial Network

검색결과 217건 처리시간 0.033초

Generative Adversarial Network를 이용한 손실된 깊이 영상 복원 (Depth Image Restoration Using Generative Adversarial Network)

  • 나준엽;심창훈;박인규
    • 방송공학회논문지
    • /
    • 제23권5호
    • /
    • pp.614-621
    • /
    • 2018
  • 본 논문에서는 generative adversarial network (GAN)을 이용한 비감독 학습을 통해 깊이 카메라로 깊이 영상을 취득할 때 발생한 손실된 부분을 복원하는 기법을 제안한다. 제안하는 기법은 3D morphable model convolutional neural network (3DMM CNN)와 large-scale CelebFaces Attribute (CelebA) 데이터 셋 그리고 FaceWarehouse 데이터 셋을 이용하여 학습용 얼굴 깊이 영상을 생성하고 deep convolutional GAN (DCGAN)의 생성자(generator)와 Wasserstein distance를 손실함수로 적용한 구별자(discriminator)를 미니맥스 게임기법을 통해 학습시킨다. 이후 학습된 생성자와 손실 부분을 복원해주기 위한 새로운 손실함수를 이용하여 또 다른 학습을 통해 최종적으로 깊이 카메라로 취득된 얼굴 깊이 영상의 손실 부분을 복원한다.

Generative Adversarial Network 를 이용한 야간 도로 영상 보정 시스템 (Night to day image translation with Generative Adversarial Network)

  • 안남현;강석주
    • 한국방송∙미디어공학회:학술대회논문집
    • /
    • 한국방송∙미디어공학회 2018년도 하계학술대회
    • /
    • pp.347-348
    • /
    • 2018
  • 본 논문에서는 야간 도로 영상을 보정하여 주간 영상으로 변환하는 알고리즘을 제안한다. 영상 변환 딥러닝 알고리즘인 Generative Adversarial Network(GAN)를 기반으로 주야간 도로 영상을 학습시켜 주야간 상호 변환이 가능한 시스템을 구현한다. 우선, 입력 영상에 대해 변환된 영상을 출력하는 generative network 를 정의한다. 또한, 변환된 영상을 다시 본래 영상으로 변환하는 inverse network 를 정의한다. Generative network 와 inverse network 를 모두 통과한 결과 영상과 본래 영상의 차 영상을 통해 손실 함수를 정의함으로써 파라미터를 목적에 맞게 학습시킬 수 있다. 또한, generative network 를 통과한 결과 영상과 목적하는 영상을 구분하는 discrimination network 를 정의하여 discrimination network 와 generative network 의 minimax two- player game 을 통해 변환된 영상이 실제 목적 영상과 유사하도록 유도한다. 제안하는 알고리즘을 적용하여 야간 도로 영상의 보정을 수행하면 주변 물체 인식이 어려운 야간 영상을 물체 인식이 용이한 주간 영상으로 변환 할 수 있다.

  • PDF

생성적 적대 신경망(GAN)을 이용한 한국어 문서에서의 문맥의존 철자오류 교정 (Context-Sensitive Spelling Error Correction Techniques in Korean Documents using Generative Adversarial Network)

  • 이정훈;권혁철
    • 한국멀티미디어학회논문지
    • /
    • 제24권10호
    • /
    • pp.1391-1402
    • /
    • 2021
  • This paper focuses use context-sensitive spelling error correction using generative adversarial network. Generative adversarial network[1] are attracting attention as they solve data generation problems that have been a challenge in the field of deep learning. In this paper, sentences are generated using word embedding information and reflected in word distribution representation. We experiment with DCGAN[2] used for the stability of learning in the existing image processing and D2GAN[3] with double discriminator. In this paper, we experimented with how the composition of generative adversarial networks and the change of learning corpus influence the context-sensitive spelling error correction In the experiment, we correction the generated word embedding information and compare the performance with the actual word embedding information.

Counterfactual image generation by disentangling data attributes with deep generative models

  • Jieon Lim;Weonyoung Joo
    • Communications for Statistical Applications and Methods
    • /
    • 제30권6호
    • /
    • pp.589-603
    • /
    • 2023
  • Deep generative models target to infer the underlying true data distribution, and it leads to a huge success in generating fake-but-realistic data. Regarding such a perspective, the data attributes can be a crucial factor in the data generation process since non-existent counterfactual samples can be generated by altering certain factors. For example, we can generate new portrait images by flipping the gender attribute or altering the hair color attributes. This paper proposes counterfactual disentangled variational autoencoder generative adversarial networks (CDVAE-GAN), specialized for data attribute level counterfactual data generation. The structure of the proposed CDVAE-GAN consists of variational autoencoders and generative adversarial networks. Specifically, we adopt a Gaussian variational autoencoder to extract low-dimensional disentangled data features and auxiliary Bernoulli latent variables to model the data attributes separately. Also, we utilize a generative adversarial network to generate data with high fidelity. By enjoying the benefits of the variational autoencoder with the additional Bernoulli latent variables and the generative adversarial network, the proposed CDVAE-GAN can control the data attributes, and it enables producing counterfactual data. Our experimental result on the CelebA dataset qualitatively shows that the generated samples from CDVAE-GAN are realistic. Also, the quantitative results support that the proposed model can produce data that can deceive other machine learning classifiers with the altered data attributes.

딥러닝 기반 교량 손상추정을 위한 Generative Adversarial Network를 이용한 가속도 데이터 생성 모델 (Generative Model of Acceleration Data for Deep Learning-based Damage Detection for Bridges Using Generative Adversarial Network)

  • 이강혁;신도형
    • 한국BIM학회 논문집
    • /
    • 제9권1호
    • /
    • pp.42-51
    • /
    • 2019
  • Maintenance of aging structures has attracted societal attention. Maintenance of the aging structure can be efficiently performed with a digital twin. In order to maintain the structure based on the digital twin, it is required to accurately detect the damage of the structure. Meanwhile, deep learning-based damage detection approaches have shown good performance for detecting damage of structures. However, in order to develop such deep learning-based damage detection approaches, it is necessary to use a large number of data before and after damage, but there is a problem that the amount of data before and after the damage is unbalanced in reality. In order to solve this problem, this study proposed a method based on Generative adversarial network, one of Generative Model, for generating acceleration data usually used for damage detection approaches. As results, it is confirmed that the acceleration data generated by the GAN has a very similar pattern to the acceleration generated by the simulation with structural analysis software. These results show that not only the pattern of the macroscopic data but also the frequency domain of the acceleration data can be reproduced. Therefore, these findings show that the GAN model can analyze complex acceleration data on its own, and it is thought that this data can help training of the deep learning-based damage detection approaches.

Deep Adversarial Residual Convolutional Neural Network for Image Generation and Classification

  • Haque, Md Foysal;Kang, Dae-Seong
    • 한국정보기술학회 영문논문지
    • /
    • 제10권1호
    • /
    • pp.111-120
    • /
    • 2020
  • Generative adversarial networks (GANs) achieved impressive performance on image generation and visual classification applications. However, adversarial networks meet difficulties in combining the generative model and unstable training process. To overcome the problem, we combined the deep residual network with upsampling convolutional layers to construct the generative network. Moreover, the study shows that image generation and classification performance become more prominent when the residual layers include on the generator. The proposed network empirically shows that the ability to generate images with higher visual accuracy provided certain amounts of additional complexity using proper regularization techniques. Experimental evaluation shows that the proposed method is superior to image generation and classification tasks.

HiGANCNN: A Hybrid Generative Adversarial Network and Convolutional Neural Network for Glaucoma Detection

  • Alsulami, Fairouz;Alseleahbi, Hind;Alsaedi, Rawan;Almaghdawi, Rasha;Alafif, Tarik;Ikram, Mohammad;Zong, Weiwei;Alzahrani, Yahya;Bawazeer, Ahmed
    • International Journal of Computer Science & Network Security
    • /
    • 제22권9호
    • /
    • pp.23-30
    • /
    • 2022
  • Glaucoma is a chronic neuropathy that affects the optic nerve which can lead to blindness. The detection and prediction of glaucoma become possible using deep neural networks. However, the detection performance relies on the availability of a large number of data. Therefore, we propose different frameworks, including a hybrid of a generative adversarial network and a convolutional neural network to automate and increase the performance of glaucoma detection. The proposed frameworks are evaluated using five public glaucoma datasets. The framework which uses a Deconvolutional Generative Adversarial Network (DCGAN) and a DenseNet pre-trained model achieves 99.6%, 99.08%, 99.4%, 98.69%, and 92.95% of classification accuracy on RIMONE, Drishti-GS, ACRIMA, ORIGA-light, and HRF datasets respectively. Based on the experimental results and evaluation, the proposed framework closely competes with the state-of-the-art methods using the five public glaucoma datasets without requiring any manually preprocessing step.

High Representation based GAN defense for Adversarial Attack

  • Sutanto, Richard Evan;Lee, Suk Ho
    • International journal of advanced smart convergence
    • /
    • 제8권1호
    • /
    • pp.141-146
    • /
    • 2019
  • These days, there are many applications using neural networks as parts of their system. On the other hand, adversarial examples have become an important issue concerining the security of neural networks. A classifier in neural networks can be fooled and make it miss-classified by adversarial examples. There are many research to encounter adversarial examples by using denoising methods. Some of them using GAN (Generative Adversarial Network) in order to remove adversarial noise from input images. By producing an image from generator network that is close enough to the original clean image, the adversarial examples effects can be reduced. However, there is a chance when adversarial noise can survive the approximation process because it is not like a normal noise. In this chance, we propose a research that utilizes high-level representation in the classifier by combining GAN network with a trained U-Net network. This approach focuses on minimizing the loss function on high representation terms, in order to minimize the difference between the high representation level of the clean data and the approximated output of the noisy data in the training dataset. Furthermore, the generated output is checked whether it shows minimum error compared to true label or not. U-Net network is trained with true label to make sure the generated output gives minimum error in the end. At last, the remaining adversarial noise that still exist after low-level approximation can be removed with the U-Net, because of the minimization on high representation terms.

정칙화 항에 기반한 WGAN의 립쉬츠 연속 안정화 기법 제안 (Technique Proposal to Stabilize Lipschitz Continuity of WGAN Based on Regularization Terms)

  • 한희일
    • 한국인터넷방송통신학회논문지
    • /
    • 제20권1호
    • /
    • pp.239-246
    • /
    • 2020
  • 최근에 제안된 WGAN(Wasserstein generative adversarial network)의 등장으로 GAN(generative adversarial network)의 고질적인 문제인 까다롭고 불안정한 학습과정이 다소 개선되기는 하였으나 여전히 수렴이 안되거나 자연스럽지 못한 출력물을 생성하는 등의 경우가 발생한다. 이러한 문제를 해결하기 위하여 본 논문에서는 분별기가 실제 데이터 확률분포를 보다 정확히 추정할 수 있도록 표본화 과정을 개선하는 동시에 분별기 함수의 립쉬츠 연속조건을 안정적으로 유지시키기 위한 알고리즘을 제안한다. 다양한 실험을 통하여 제안 기법의 특성을 분석하고 성능을 확인한다.