과제정보
이 논문은 부경대학교 자율창의학술연구비(2021년)에 의하여 연구되었음.
참고문헌
- I.J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A.C. Courville, Y. Bengio, "Generative adversarial nets," In Advances in Neural Information Processing Systems, pp.2672-2680, 2014. doi: https://doi.org/10.48550/arXiv.1406.2661
- T. Karras, T. Aila, S. Laine, J. Lehtinen, "Progressive Growing of GANs for Improved Quality, Stability, and Variation," arXiv preprint arXiv:1710.10196, 2017. doi: https://doi.org/10.48550/arXiv.1710.10196
- X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, P. Abbeel, "InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets," 30th Conference on Neural Information Processing Systems, NIPS, 2016. doi: https://doi.org/10.48550/arXiv.1606.03657
- J. Y. Zhu, T. Park, P. Isola, A. A. Efros, "Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks," Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp.2223-2232, 2017 doi: http://dx.doi.org/10.1109/ICCV.2017.244
- Moon, C. K., Uh, Y. J., Byun, H. R., "Image Restoration using GAN," JOURNAL OF BROADCAST ENGINEERING, pp.503-510, 2018 doi: https://doi.org/10.5909/JBE.2018.23.4.503
- Nah, J. Y., Sim, C. H., Park, I. K., "Depth Image Restoration Using Generative Adversarial Network," JOURNAL OF BROADCAST ENGINEERING, pp.614-621, 2018 doi: https://doi.org/10.5909/JBE.2018.23.5.614
- Kim, k. h., Kong, K. B., Kang, S. J., "A Review on Deep Learning-based Image Outpainting," JOURNAL OF BROADCAST ENGINEERING, pp.61-69, 2021 doi: https://doi.org/10.5909/JBE.2021.26.1.61
- M. Arjovsky, S. Chintala, L. Bottou, "Wasserstein Generative Adversarial Networks," Proceedings of the 34th International Conference on Machine Learning, pp. 214-223, 2017. doi: https://doi.org/10.48550/arXiv.1701.07875
- T. Miyato, T. Kataoka, M. Koyama, Y. Yoshida, "Spectral normalization for generative adversarial networks," arXiv preprint arXiv:1802.05957, 2018. doi: https://doi.org/10.48550/arXiv.1802.05957
- T.R. Shaham, T. Dekel, T. Michaeli, "Singan: Learning a generative model from a single natural image," Proceedings of the IEEE International Conference on Computer Vision, pp.4570-4580, 2019. doi: https://doi.org/10.48550/arXiv.1905.01164
- T.Hinz, M. Fisher, O. Wang, S. Wermter, "Improved techniques for training single-image gans," In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp.1300-1309, 2021. doi: http://dx.doi.org/10.1109/WACV48630.2021.00134
- A. Shocher, S. Bagon, P. Isola, M. Irani, "Ingan: Capturing and remapping the" dna" of a natural image," In IEEE International Conference on Computer Vision (ICCV) 2019. doi: http://dx.doi.org/10.1109/ICCV.2019.00459
- Y. Vinker, E. Horwitz, N. Zabari, "Image Shape Manipulation from a Single Augmented Training Sample," Proceedings of the ICCV, 2021. doi: http://dx.doi.org/10.1109/ICCV48922.2021.01351
- V. Sushko, J. Gall, A. Khoreva, "One-shot gan: Learning to generate samples from single images and videos," in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.2596-2600. 2021. doi: http://dx.doi.org/10.1109/CVPRW53098.2021.00293
- X. Mao, Q. Li, H. Xie, R.Y.K. Lau, Z. Wang, "Least squares generative adversarial networks," Proceedings of the ICCV, 2016. doi: http://dx.doi.org/10.1109/ICCV.2017.304
- D. Berthelot, T. Schumm, L. Metz, "BEGAN: Boundary Equilibrium Generative Adversarial Networks," arXiv preprint arXiv:1703.10717, 2017. doi: https://doi.org/10.48550/arXiv.1703.10717
- A. Radford, L. Metz, S. Chintala, "Unsupervised representation learning with deep convolutional generative adversarial networks," arXiv preprint arXiv:1511.06434, 2015. doi: https://doi.org/10.48550/arXiv.1511.06434
- H. Zhang, I. Goodfellow, D. Metaxas, "Self-Attention Generative Adversarial Networks," Proceedings of the 36th International Conference on Machine Learning, pp.7354-7363, 2019. doi: https://doi.org/10.48550/arXiv.1805.08318
- G. Donato, S. Belongie, "Approximate thin plate spline mappings," In European conference on computer vision, pp.21-31, 2002. doi: http://dx.doi.org/10.1007/3-540-47977-5_2
- M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, S. Hochreiter, "Gans trained by a two time-scale update rule converge to a local nash equilibrium," In Advances in Neural Information Processing Systems, pp.6626-6637, 2017. doi: https://dl.acm.org/doi/10.5555/3295222.3295408
- J. Lin, Y. Pang, Y. Xia, Z. Chen, J. Luo, "Tuigan: Learning versatile image-to-image translation with two unpaired images," European Conference on Computer Vision. pp.18-35. 2020. doi: http://dx.doi.org/10.1007/978-3-030-58548-8_2
- R. Zhang, P.Isola, A. A. Efros, E. Shechtman, O. Wang, "The unreasonable effectiveness of deep features as a perceptual metric," arXiv preprint arXiv:1801.03924, 2018. doi: http://dx.doi.org/10.1109/CVPR.2018.00068
- T. C. Wang, M. Y. Liu, J. Y. Zhu, A. Tao, J. Kautz, B. Catanzaro, "High-resolution image synthesis and semantic manipulation with conditional gans," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.8798-8807, 2018 doi: http://dx.doi.org/10.1109/CVPR.2018.00917
- https://github.com/tamarott/SinGAN
- https://github.com/tohinz/ConSinGAN
- https://github.com/assafshocher/InGAN
- https://github.com/eliahuhorwitz/DeepSIM