參考文獻 |
[1] I. Goodfellow, J. Pouget-Abadie, M. Mirza, et al., “Generative Adversarial Nets,” in
Advances in Neural Information Processing Systems, Z. Ghahramani, M. Welling, C.
Cortes, N. Lawrence, and K. Q. Weinberger, Eds., vol. 27, Curran Associates, Inc., 2014.
[2] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks,” in 2017 IEEE International Conference on
Computer Vision (ICCV), Oct. 2017, pp. 2242–2251.
[3] T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive Growing of GANs for Improved Quality, Stability, and Variation,” in International Conference on Learning Representations, Feb. 2018.
[4] T. Karras, S. Laine, and T. Aila, “A Style-Based Generator Architecture for Generative
Adversarial Networks,” in 2019 IEEE/CVF Conference on Computer Vision and Pattern
Recognition (CVPR), Jun. 2019, pp. 4396–4405.
[5] T. Karras, S. Laine, M. Aittala, J. Hellsten, J. Lehtinen, and T. Aila, “Analyzing and Improving the Image Quality of StyleGAN,” in 2020 IEEE/CVF Conference on Computer
Vision and Pattern Recognition (CVPR), Jun. 2020, pp. 8107–8116.
[6] T. Karras, M. Aittala, J. Hellsten, S. Laine, J. Lehtinen, and T. Aila, “Training Generative Adversarial Networks with Limited Data,” in Thirty-Fourth Conference on Neural
Information Processing Systems, 2020.
[7] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE conference on computer vision
and pattern recognition, 2017, pp. 1125–1134.
[8] B. Albahar, J. Lu, J. Yang, Z. Shu, E. Shechtman, and J.-B. Huang, “Pose with style:
Detail-preserving pose-guided image synthesis with conditional stylegan,” ACM Transactions on Graphics (TOG), vol. 40, no. 6, pp. 1–11, 2021.
[9] Z. Cao, T. Simon, S.-E. Wei, and Y. Sheikh, “Realtime Multi-person 2D Pose Estimation
Using Part Affinity Fields,” in 2017 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), Jul. 2017, pp. 1302–1310.
[10] Z. Cao, G. Hidalgo, T. Simon, S.-E. Wei, and Y. Sheikh, “OpenPose: Realtime MultiPerson 2D Pose Estimation Using Part Affinity Fields,” IEEE Transactions on Pattern
Analysis and Machine Intelligence, vol. 43, no. 1, pp. 172–186, Jan. 2019.
[11] H. Fang, S. Xie, Y.-W. Tai, and C. Lu, “RMPE: Regional Multi-person Pose Estimation,”
2017 IEEE International Conference on Computer Vision (ICCV), 2017.
[12] K. He, G. Gkioxari, P. Dollár, and R. Girshick, “Mask R-CNN,” in 2017 IEEE International Conference on Computer Vision (ICCV), Oct. 2017, pp. 2980–2988.
[13] Wikipedia. “UV mapping.” (Jun. 23, 2022), [Online]. Available: https://en.wikipedia.
org/wiki/UV_mapping (visited on 07/04/2022).
[14] R. A. Güler, N. Neverova, and I. Kokkinos, “DensePose: Dense Human Pose Estimation
in the Wild,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2018, pp. 7297–7306.
[15] R. A. Güler, G. Trigeorgis, E. Antonakos, P. Snape, S. Zafeiriou, and I. Kokkinos, “DenseReg:
Fully Convolutional Dense Shape Regression In-the-Wild,” in 2017 IEEE Conference on
Computer Vision and Pattern Recognition (CVPR), Jul. 2017, pp. 2614–2623.
[16] S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” IEEE Transactions on Pattern Analysis and
Machine Intelligence, vol. 39, no. 6, pp. 1137–1149, Jun. 2017.
[17] N.-C. Lee et al., “應用生成對抗網路於嬰兒骨架偵測與早產兒整體動作指標分析,”
M.S. thesis, National Central University, 2020.
[18] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint
arXiv:1411.1784, 2014.
[19] C. Chan, S. Ginosar, T. Zhou, and A. A. Efros, “Everybody dance now,” in Proceedings
of the IEEE/CVF international conference on computer vision, 2019, pp. 5933–5942.
[20] L. Ma, X. Jia, Q. Sun, B. Schiele, T. Tuytelaars, and L. Van Gool, “Pose guided person
image generation,” Advances in neural information processing systems, vol. 30, 2017.
[21] G. Yildirim, N. Jetchev, R. Vollgraf, and U. Bergmann, “Generating high-resolution fashion model images wearing custom outfits,” in Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, 2019, pp. 0–0.
[22] Y. Men, Y. Mao, Y. Jiang, W.-Y. Ma, and Z. Lian, “Controllable person image synthesis
with attribute-decomposed gan,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 5084–5093.
[23] K. Sarkar, V. Golyanik, L. Liu, and C. Theobalt, “Style and pose control for image synthesis of humans from a single monocular view,” arXiv preprint arXiv:2102.11263, 2021.
[24] E. Lu, F. Cole, T. Dekel, et al., “Layered neural rendering for retiming people in video,”
arXiv preprint arXiv:2009.07833, 2020.43
[25] R. Abdal, Y. Qin, and P. Wonka, “Image2stylegan: How to embed images into the stylegan latent space?” In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 4432–4441.
[26] A. Tewari, M. Elgharib, G. Bharaj, et al., “Stylerig: Rigging stylegan for 3d control over
portrait images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, 2020, pp. 6142–6151.
[27] B. Egger, W. A. Smith, A. Tewari, et al., “3d morphable face models—past, present, and
future,” ACM Transactions on Graphics (TOG), vol. 39, no. 5, pp. 1–38, 2020.
[28] D. Castro, S. Hickson, P. Sangkloy, et al., “Let’s dance: Learning from online dance
videos,” arXiv preprint arXiv:1801.07388, 2018.
[29] M. Andriluka, L. Pishchulin, P. Gehler, and B. Schiele, “2d human pose estimation: New
benchmark and state of the art analysis,” in Proceedings of the IEEE Conference on
computer Vision and Pattern Recognition, 2014, pp. 3686–3693.
[30] Z. Liu, P. Luo, S. Qiu, X. Wang, and X. Tang, “Deepfashion: Powering robust clothes
recognition and retrieval with rich annotations,” in Proceedings of the IEEE conference
on computer vision and pattern recognition, 2016, pp. 1096–1104.
[31] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and
computer-assisted intervention, Springer, 2015, pp. 234–241.
[32] P. Zablotskaia, A. Siarohin, B. Zhao, and L. Sigal, “Dwnet: Dense warp-based network
for pose-guided human video generation,” arXiv preprint arXiv:1910.09139, 2019.
[33] L. Ma, Q. Sun, S. Georgoulis, L. Van Gool, B. Schiele, and M. Fritz, “Disentangled
person image generation,” in Proceedings of the IEEE Conference on Computer Vision
and Pattern Recognition, 2018, pp. 99–108. |