• 제목/요약/키워드: conditional-GANs

검색결과 9건 처리시간 0.022초

cGANs 기반 3D 포인트 클라우드 데이터의 실시간 전송 기법 (Real-time transmission of 3G point cloud data based on cGANs)

  • Shin, Kwang-Seong;Shin, Seong-Yoon
    • 한국정보통신학회논문지
    • /
    • 제23권11호
    • /
    • pp.1482-1484
    • /
    • 2019
  • We present a method for transmitting 3D object information in real time in a telepresence system. Three-dimensional object information consists of a large amount of point cloud data, which requires high performance computing power and ultra-wideband network transmission environment to process and transmit such a large amount of data in real time. In this paper, multiple users can transmit object motion and facial expression information in real time even in small network bands by using GANs (Generative Adversarial Networks), a non-supervised learning machine learning algorithm, for real-time transmission of 3D point cloud data. In particular, we propose the creation of an object similar to the original using only the feature information of 3D objects using conditional GANs.

다중 스케일 그라디언트 조건부 적대적 생성 신경망을 활용한 문장 기반 영상 생성 기법 (Text-to-Face Generation Using Multi-Scale Gradients Conditional Generative Adversarial Networks)

  • ;;추현승
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2021년도 추계학술발표대회
    • /
    • pp.764-767
    • /
    • 2021
  • While Generative Adversarial Networks (GANs) have seen huge success in image synthesis tasks, synthesizing high-quality images from text descriptions is a challenging problem in computer vision. This paper proposes a method named Text-to-Face Generation Using Multi-Scale Gradients for Conditional Generative Adversarial Networks (T2F-MSGGANs) that combines GANs and a natural language processing model to create human faces has features found in the input text. The proposed method addresses two problems of GANs: model collapse and training instability by investigating how gradients at multiple scales can be used to generate high-resolution images. We show that T2F-MSGGANs converge stably and generate good-quality images.

광학 영상의 구름 제거를 위한 조건부 생성적 적대 신경망과 회귀 기반 보정의 결합 (Combining Conditional Generative Adversarial Network and Regression-based Calibration for Cloud Removal of Optical Imagery)

  • 곽근호;박소연;박노욱
    • 대한원격탐사학회지
    • /
    • 제38권6_1호
    • /
    • pp.1357-1369
    • /
    • 2022
  • 구름 제거는 식생 모니터링, 변화 탐지 등과 같은 광학 영상이 필요한 모든 작업에서 필수적인 영상 처리 과정이다. 이 논문에서는 조건부 생성적 적대 신경망(conditional generative adversarial networks, cGANs)과 회귀 기반 보정을 결합하여 구름이 없는 시계열 광학 영상 세트를 구성하는 2단계의 구름 제거 기법을 제안하였다. 첫 번째 단계에서는 광학 영상과 synthetic aperture radar 영상 간 정량적 관계를 이용하는 cGANs을 이용하여 초기 예측 결과를 생성한다. 두 번째 단계에서는 구름이 아닌 영역에서 예측 결과와 실제 값과의 관계를 random forest 기반 회귀 모델링을 통해 정량화한 후에 cGANs 기반 예측 결과를 보정한다. 제안 기법은 김제의 벼 재배지에서 Sentinel-2 영상과 COSMO-SkyMed 영상을 이용한 구름 제거 실험을 통해 적용 가능성을 평가하였다. cGAN 모델은 구름 영역에서 지표면 상태의 급격한 변화가 발생하는 논 재배지를 대상으로 반사율 값을 효과적으로 예측할 수 있었다. 또한 두 번째 단계의 회귀 기반 보정은 예측 대상 영상에서 시간적으로 떨어진 보조 영상을 이용하는 회귀 기반 구름 제거 기법에 비해 예측 정확도를 향상시킬 수 있었다. 이러한 실험 결과는 구름이 없는 광학 영상을 환경 모니터링에 이용할 수 없는 경우 제안된 방법이 구름 오염 지역을 복원하는데 효과적으로 적용될 수 있음을 나타낸다.

cGANs(Conditional Generative Adversarial Networks) 기반 3차원 객체의 임의 재생 기법 연구 (A Study on Random Reconstruction Method of 3-D Objects Based on Conditional Generative Adversarial Networks (cGANs))

  • 신광성;신성윤
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2019년도 춘계학술대회
    • /
    • pp.157-159
    • /
    • 2019
  • 홀로그램 기술은 3차원 객체의 생성, 전송, 재생 관점에서 기술 개발이 활발히 수행되고 있지만 현재 여러 가지 한계로 인하여 답보상태에 머물러 있다. VR, AR을 넘어 새로운 기술의 요구에 부합하기 위해 중간 단계로 유사홀로그램 시장이 성장하고 있는 추세다. 홀로그램의 기술의 핵심은 point cloud 형태의 방대한 3차원 데이터를 생성하고 그 방대한 데이터를 통신망을 통해 실시간으로 전송하여 목적지에서 원본과 같이 재생하는 것이다. 본 연구에서는 방대한 3차원 데이터를 실시간으로 전송하기 위한 방법으로 생성된 3차원 객체 정보의 특징점을 전송하여 목적지에서 원본과 비슷한 형태의 객체로 재생하는 방법에 대해 연구한다.

  • PDF

다수 화자 한국어 음성 변환 실험 (Many-to-many voice conversion experiments using a Korean speech corpus)

  • 육동석;서형진;고봉구;유인철
    • 한국음향학회지
    • /
    • 제41권3호
    • /
    • pp.351-358
    • /
    • 2022
  • 심층 생성 모델의 일종인 Generative Adversarial Network(GAN)과 Variational AutoEncoder(VAE)는 비병렬 학습 데이터를 사용한 음성 변환에 새로운 방법론을 제시하고 있다. 특히, Conditional Cycle-Consistent Generative Adversarial Network(CC-GAN)과 Cycle-Consistent Variational AutoEncoder(CycleVAE)는 다수 화자 사이의 음성 변환에 우수한 성능을 보이고 있다. 그러나, CC-GAN과 CycleVAE는 비교적 적은 수의 화자를 대상으로 연구가 진행되어왔다. 본 논문에서는 100 명의 한국어 화자 데이터를 사용하여 CC-GAN과 CycleVAE의 음성 변환 성능과 확장 가능성을 실험적으로 분석하였다. 실험 결과 소규모 화자의 경우 CC-GAN이 Mel-Cepstral Distortion(MCD) 기준으로 4.5 % 우수한 성능을 보이지만 대규모 화자의 경우 CycleVAE가 제한된 학습 시간 안에 12.7 % 우수한 성능을 보였다.

Solar farside magnetograms from deep learning analysis of STEREO/EUVI data

  • Kim, Taeyoung;Park, Eunsu;Lee, Harim;Moon, Yong-Jae;Bae, Sung-Ho;Lim, Daye;Jang, Soojeong;Kim, Lokwon;Cho, Il-Hyun;Choi, Myungjin;Cho, Kyung-Suk
    • 천문학회보
    • /
    • 제44권1호
    • /
    • pp.51.3-51.3
    • /
    • 2019
  • Solar magnetograms are important for studying solar activity and predicting space weather disturbances1. Farside magnetograms can be constructed from local helioseismology without any farside data2-4, but their quality is lower than that of typical frontside magnetograms. Here we generate farside solar magnetograms from STEREO/Extreme UltraViolet Imager (EUVI) $304-{\AA}$ images using a deep learning model based on conditional generative adversarial networks (cGANs). We train the model using pairs of Solar Dynamics Observatory (SDO)/Atmospheric Imaging Assembly (AIA) $304-{\AA}$ images and SDO/Helioseismic and Magnetic Imager (HMI) magnetograms taken from 2011 to 2017 except for September and October each year. We evaluate the model by comparing pairs of SDO/HMI magnetograms and cGAN-generated magnetograms in September and October. Our method successfully generates frontside solar magnetograms from SDO/AIA $304-{\AA}$ images and these are similar to those of the SDO/HMI, with Hale-patterned active regions being well replicated. Thus we can monitor the temporal evolution of magnetic fields from the farside to the frontside of the Sun using SDO/HMI and farside magnetograms generated by our model when farside extreme-ultraviolet data are available. This study presents an application of image-to-image translation based on cGANs to scientific data.

  • PDF

Application of Deep Learning to Solar Data: 3. Generation of Solar images from Galileo sunspot drawings

  • Lee, Harim;Moon, Yong-Jae;Park, Eunsu;Jeong, Hyunjin;Kim, Taeyoung;Shin, Gyungin
    • 천문학회보
    • /
    • 제44권1호
    • /
    • pp.81.2-81.2
    • /
    • 2019
  • We develop an image-to-image translation model, which is a popular deep learning method based on conditional Generative Adversarial Networks (cGANs), to generate solar magnetograms and EUV images from sunspot drawings. For this, we train the model using pairs of sunspot drawings from Mount Wilson Observatory (MWO) and their corresponding SDO/HMI magnetograms and SDO/AIA EUV images (512 by 512) from January 2012 to September 2014. We test the model by comparing pairs of actual SDO images (magnetogram and EUV images) and the corresponding AI-generated ones from October to December in 2014. Our results show that bipolar structures and coronal loop structures of AI-generated images are consistent with those of the original ones. We find that their unsigned magnetic fluxes well correlate with those of the original ones with a good correlation coefficient of 0.86. We also obtain pixel-to-pixel correlations EUV images and AI-generated ones. The average correlations of 92 test samples for several SDO lines are very good: 0.88 for AIA 211, 0.87 for AIA 1600 and 0.93 for AIA 1700. These facts imply that AI-generated EUV images quite similar to AIA ones. Applying this model to the Galileo sunspot drawings in 1612, we generate HMI-like magnetograms and AIA-like EUV images of the sunspots. This application will be used to generate solar images using historical sunspot drawings.

  • PDF

Application of Deep Learning to Solar Data: 1. Overview

  • Moon, Yong-Jae;Park, Eunsu;Kim, Taeyoung;Lee, Harim;Shin, Gyungin;Kim, Kimoon;Shin, Seulki;Yi, Kangwoo
    • 천문학회보
    • /
    • 제44권1호
    • /
    • pp.51.2-51.2
    • /
    • 2019
  • Multi-wavelength observations become very popular in astronomy. Even though there are some correlations among different sensor images, it is not easy to translate from one to the other one. In this study, we apply a deep learning method for image-to-image translation, based on conditional generative adversarial networks (cGANs), to solar images. To examine the validity of the method for scientific data, we consider several different types of pairs: (1) Generation of SDO/EUV images from SDO/HMI magnetograms, (2) Generation of backside magnetograms from STEREO/EUVI images, (3) Generation of EUV & X-ray images from Carrington sunspot drawing, and (4) Generation of solar magnetograms from Ca II images. It is very impressive that AI-generated ones are quite consistent with actual ones. In addition, we apply the convolution neural network to the forecast of solar flares and find that our method is better than the conventional method. Our study also shows that the forecast of solar proton flux profiles using Long and Short Term Memory method is better than the autoregressive method. We will discuss several applications of these methodologies for scientific research.

  • PDF

Image Translation of SDO/AIA Multi-Channel Solar UV Images into Another Single-Channel Image by Deep Learning

  • Lim, Daye;Moon, Yong-Jae;Park, Eunsu;Lee, Jin-Yi
    • 천문학회보
    • /
    • 제44권2호
    • /
    • pp.42.3-42.3
    • /
    • 2019
  • We translate Solar Dynamics Observatory/Atmospheric Imaging Assembly (AIA) ultraviolet (UV) multi-channel images into another UV single-channel image using a deep learning algorithm based on conditional generative adversarial networks (cGANs). The base input channel, which has the highest correlation coefficient (CC) between UV channels of AIA, is 193 Å. To complement this channel, we choose two channels, 1600 and 304 Å, which represent upper photosphere and chromosphere, respectively. Input channels for three models are single (193 Å), dual (193+1600 Å), and triple (193+1600+304 Å), respectively. Quantitative comparisons are made for test data sets. Main results from this study are as follows. First, the single model successfully produce other coronal channel images but less successful for chromospheric channel (304 Å) and much less successful for two photospheric channels (1600 and 1700 Å). Second, the dual model shows a noticeable improvement of the CC between the model outputs and Ground truths for 1700 Å. Third, the triple model can generate all other channel images with relatively high CCs larger than 0.89. Our results show a possibility that if three channels from photosphere, chromosphere, and corona are selected, other multi-channel images could be generated by deep learning. We expect that this investigation will be a complementary tool to choose a few UV channels for future solar small and/or deep space missions.

  • PDF