skip to main content
survey

Controllable Data Generation by Deep Learning: A Review

Published:25 April 2024Publication History
Skip Abstract Section

Abstract

Designing and generating new data under targeted properties has been attracting various critical applications such as molecule design, image editing and speech synthesis. Traditional hand-crafted approaches heavily rely on expertise experience and intensive human efforts, yet still suffer from the insufficiency of scientific knowledge and low throughput to support effective and efficient data generation. Recently, the advancement of deep learning has created the opportunity for expressive methods to learn the underlying representation and properties of data. Such capability provides new ways of determining the mutual relationship between the structural patterns and functional properties of the data and leveraging such relationships to generate structural data, given the desired properties. This article is a systematic review that explains this promising research area, commonly known as controllable deep data generation. First, the article raises the potential challenges and provides preliminaries. Then the article formally defines controllable deep data generation, proposes a taxonomy on various techniques and summarizes the evaluation metrics in this specific domain. After that, the article introduces exciting applications of controllable deep data generation, experimentally analyzes and compares existing works. Finally, this article highlights the promising future directions of controllable deep data generation and identifies five potential challenges.

Skip Supplemental Material Section

Supplemental Material

REFERENCES

  1. [1] You Jiaxuan, Liu Bowen, et al. 2018. Graph convolutional policy network for goal-directed molecular graph generation. Conference on Neural Information Processing Systems 31 (2018).Google ScholarGoogle Scholar
  2. [2] Jin Wengong, Barzilay Regina, and Jaakkola Tommi. 2018. Junction tree variational autoencoder for molecular graph generation. In International Conference on Machine Learning. PMLR, 23232332.Google ScholarGoogle Scholar
  3. [3] Cao Nicola De and Kipf Thomas. 2018. MolGAN: An implicit generative model for small molecular graphs. International Conference on Machine Learning 2018 Workshop on Theoretical Foundations and Applications of Deep Generative Models (2018).Google ScholarGoogle Scholar
  4. [4] Gregor Karol, Danihelka Ivo, et al. 2015. Draw: A recurrent neural network for image generation. In International Conference on Machine Learning. PMLR, 14621471.Google ScholarGoogle Scholar
  5. [5] Mirza Mehdi and Osindero Simon. 2014. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014).Google ScholarGoogle Scholar
  6. [6] Liu Rui, Liu Yu, et al. 2019. Conditional adversarial generative flow for controllable image synthesis. In Conference on Computer Vision and Pattern Recognition. 79928001.Google ScholarGoogle ScholarCross RefCross Ref
  7. [7] Guo Jiaxian, Lu Sidi, et al. 2018. Long text generation via adversarial training with leaked information. In AAAI Conference on Artificial Intelligence, Vol. 32.Google ScholarGoogle ScholarCross RefCross Ref
  8. [8] Tambwekar Pradyumna, Dhuliawala Murtaza, et al. 2019. Controllable neural story plot generation via reward shaping. In International Joint Conferences on Artificial Intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Schröder Marc. 2001. Emotional speech synthesis: A review. In European Conference on Speech Communication and Technology. Citeseer.Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Yang Zhongliang, Du Xingjian, et al. 2018. AAG-Stega: Automatic audio generation-based steganography. arXiv preprint arXiv:1809.03463 (2018).Google ScholarGoogle Scholar
  11. [11] Habib Raza, Mariooryad Soroosh, et al. 2020. Semi-supervised generative modeling for controllable speech synthesis. In International Conference on Learning Representations. https://openreview.net/forum?id=rJeqeCEtvHGoogle ScholarGoogle Scholar
  12. [12] Lippow Shaun M. and Tidor Bruce. 2007. Progress in computational protein design. Current Opinion in Biotechnology 18, 4 (2007), 305311.Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Vourkas Ioannis and Sirakoulis Georgios Ch.. 2016. Emerging memristor-based logic circuit design approaches: A review. IEEE Circuits and Systems Magazine 16, 3 (2016), 1530.Google ScholarGoogle ScholarCross RefCross Ref
  14. [14] Yang Hai and Bell Michael G. H.. 1998. Models and algorithms for road network design: A review and some new developments. Transport Reviews 18, 3 (1998), 257278.Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Fischer-Hübner Simone. 2001. IT-security and Privacy: Design and Use of Privacy-enhancing Security Mechanisms. Springer.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Jensen Jan H.. 2019. A graph-based genetic algorithm and generative model/Monte Carlo tree search for the exploration of chemical space. Chemical Science 10, 12 (2019), 35673572.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Polishchuk Pavel G., Madzhidov Timur I., and Varnek Alexandre. 2013. Estimation of the size of drug-like chemical space based on GDB-17 data. Journal of Computer-aided Molecular Design 27, 8 (2013), 675679.Google ScholarGoogle ScholarCross RefCross Ref
  18. [18] Uyemura John P.. 1999. CMOS Logic Circuit Design. Springer Science & Business Media.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. [19] Chen Qiao, Wang Xiaoping, et al. 2016. A logic circuit design for perfecting memristor-based material implication. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 36, 2 (2016), 279284.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Huang Po-Ssu, Boyken Scott E., and Baker David. 2016. The coming of age of de novo protein design. Nature 537, 7620 (2016), 320327.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Farahani Reza Zanjirani, Miandoabchi Elnaz, et al. 2013. A review of urban transportation network design problems. European Journal of Operational Research (2013).Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Kan Xuan, Cui Hejie, et al. 2022. FBNETGEN: Task-aware GNN-based fMRI analysis via functional brain network generation. In Medical Imaging with Deep Learning. https://openreview.net/forum?id=oWFphg2IKonGoogle ScholarGoogle Scholar
  23. [23] Ling Chen, Yang Carl, and Zhao Liang. 2021. Deep generation of heterogeneous networks. In International Conference on Data Mining. IEEE, 379388.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Shrestha Ajay and Mahmood Ausif. 2019. Review of deep learning algorithms and architectures. IEEE Access 7 (2019), 5304053065.Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Pouyanfar Samira, Sadiq Saad, et al. 2018. A survey on deep learning: Algorithms, techniques, and applications. Comput. Surveys 51, 5 (2018), 136.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Pamina J. and Raja Beschi. 2019. Survey on deep learning algorithms. International Journal of Emerging Technology and Innovative Engineering 5, 1 (2019).Google ScholarGoogle Scholar
  27. [27] Wang Shiyu, Guo Xiaojie, Lin Xuanyang, Pan Bo, Du Yuanqi, Wang Yinkai, Ye Yanfang, Petersen Ashley Ann, Leitgeb Austin, AlKhalifa Saleh, Minbiole Kevin, Wuest Bill, Shehu Amarda, and Zhao Liang. 2022. Multi-objective Deep Data Generation with Correlated Property Control. (2022). arxiv:cs.LG/2210.01796Google ScholarGoogle Scholar
  28. [28] Zhang Zheng and Zhao Liang. 2021. Representation learning on spatial networks. Advances in Neural Information Processing Systems 34 (2021), 23032318.Google ScholarGoogle Scholar
  29. [29] Muller Alex T., Hiss Jan A., and Schneider Gisbert. 2018. Recurrent neural network model for constructive peptide design. Journal of Chemical Information and Modeling 58, 2 (2018), 472479.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Ingraham John, Garg Vikas, et al. 2019. Generative models for graph-based protein design. Conference on Neural Information Processing Systems 32 (2019).Google ScholarGoogle Scholar
  31. [31] Anand Namrata, Eguchi Raphael, et al. 2022. Protein sequence design with a learned potential. Nature Communications 13, 1 (2022), 111.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Hossain MD Zakir, Sohel Ferdous, et al. 2019. A comprehensive survey of deep learning for image captioning. Comput. Surveys 51, 6 (2019), 136.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. [33] Xie Yutong, Shi Chence, et al. 2021. MARS: Markov molecular sampling for multi-objective drug discovery. In International Conference on Learning Representations. https://openreview.net/forum?id=kHSu4ebxFXYGoogle ScholarGoogle Scholar
  34. [34] Lambrinidis George and Tsantili-Kakoulidou Anna. 2021. Multi-objective optimization methods in novel drug design. Expert Opinion on Drug Discovery 16, 6 (2021), 647658.Google ScholarGoogle ScholarCross RefCross Ref
  35. [35] Croft William L., Sack Jörg-Rüdiger, and Shi Wei. 2022. Differentially private facial obfuscation via generative adversarial networks. Future Generation Computer Systems 129 (2022), 358379.Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Li Mao, Lv Jiancheng, et al. 2020. An abstract painting generation method based on deep generative model. Neural Processing Letters 52, 2 (2020), 949960.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. [37] Mufti Adeel, Antonelli Biagio, and Monello Julius. 2020. Conditional GANs for painting generation. In International Conference on Machine Vision. SPIE.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Tan Hao Hao, Luo Yin-Jyun, and Herremans Dorien. 2020. Generative modelling for controllable audio synthesis of piano performance. In International Conference on Machine Learning Workshop on Machine Learning for Music Discovery Workshop (ML4MD), Extended Abstract.Google ScholarGoogle Scholar
  39. [39] Tits Noé, Wang Fengna, et al. 2019. Visualization and interpretation of latent spaces for controlling expressive speech synthesis through audio analysis. In Proc. Interspeech 2019. 44754479.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Buffet-Bataillon Sylvie, Branger Bernard, et al. 2011. Effect of higher minimum inhibitory concentrations of quaternary ammonium compounds in clinical E. coli isolates on antibiotic susceptibilities and clinical outcomes. Journal of Hospital Infection 79, 2 (2011), 141146.Google ScholarGoogle ScholarCross RefCross Ref
  41. [41] Deng Chaorui, Ding Ning, et al. 2020. Length-controllable image captioning. In European Conference on Computer Vision. Springer, 712729.Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. [42] Guo Longteng, Liu Jing, et al. 2019. MSCap: Multi-style image captioning with unpaired stylized text. In Conference on Computer Vision and Pattern Recognition. 42044213.Google ScholarGoogle ScholarCross RefCross Ref
  43. [43] Guo Xiaojie, Du Yuanqi, and Zhao Liang. 2021. Property controllable variational autoencoder via invertible mutual dependence. In International Conference on Learning Representations. https://openreview.net/forum?id=tYxG_OMs9WEGoogle ScholarGoogle Scholar
  44. [44] Chen Li-Wei and Rudnicky Alexander. 2022. Fine-grained style control in Transformer-based text-to-speech synthesis. In IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 79077911.Google ScholarGoogle ScholarCross RefCross Ref
  45. [45] Hu Zhiting, Yang Zichao, et al. 2017. Toward controlled generation of text. In International Conference on Machine Learning. PMLR, 15871596.Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. [46] Tolls Johannes, Dijk John van, et al. 2002. Aqueous solubility- molecular size relationships: A mechanistic case study using C10-to C19-alkanes. The Journal of Physical Chemistry A 106, 11 (2002), 27602765.Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Hu Zhiting and Li Li Erran. 2021. A causal lens for controllable text generation. In Conference on Neural Information Processing Systems, Beygelzimer A., Dauphin Y., Liang P., and Vaughan J. Wortman (Eds.). https://openreview.net/forum?id=kAm9By0R5MEGoogle ScholarGoogle Scholar
  48. [48] Du Yuanqi, Guo Xiaojie, et al. 2022. Interpretable molecular graph generation via monotonic constraints. In SIAM International Conference on Data Mining. SIAM, 7381.Google ScholarGoogle ScholarCross RefCross Ref
  49. [49] Zhang Jianzhi. 2000. Protein-length distributions for the three domains of life. Trends in Genetics 16, 3 (2000), 107109.Google ScholarGoogle ScholarCross RefCross Ref
  50. [50] Djuke Veldhuis. 2011. Tree-like giant is largest molecule ever made. New Scientist 209, 2795 (2011), 17. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  51. [51] Kirkpatrick Peter and Ellis Clare. 2004. Chemical space. Nature 432, 7019 (2004), 823824.Google ScholarGoogle ScholarCross RefCross Ref
  52. [52] Landrum Greg et al. 2013. RDKit: A software suite for cheminformatics, computational chemistry, and predictive modeling. Greg Landrum (2013).Google ScholarGoogle Scholar
  53. [53] Dong Shi, Wang Ping, and Abbas Khushnood. 2021. A survey on deep learning and its applications. Computer Science Review 40 (2021), 100379.Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. [54] Alom Md. Zahangir, Taha Tarek M., et al. 2019. A state-of-the-art survey on deep learning theory and architectures. Electronics 8, 3 (2019), 292.Google ScholarGoogle ScholarCross RefCross Ref
  55. [55] Deng Li. 2014. A tutorial survey of architectures, algorithms, and applications for deep learning. APSIPA Transactions on Signal and Information Processing 3 (2014).Google ScholarGoogle Scholar
  56. [56] Hatcher William Grant and Yu Wei. 2018. A survey of deep learning: Platforms, applications and emerging research trends. IEEE Access 6 (2018), 2441124432.Google ScholarGoogle ScholarCross RefCross Ref
  57. [57] Liu Weibo, Wang Zidong, et al. 2017. A survey of deep neural network architectures and their applications. Neurocomputing 234 (2017), 1126.Google ScholarGoogle ScholarCross RefCross Ref
  58. [58] Dargan Shaveta, Kumar Munish, et al. 2020. A survey of deep learning and its applications: A new paradigm to machine learning. Archives of Computational Methods in Engineering 27, 4 (2020), 10711092.Google ScholarGoogle ScholarCross RefCross Ref
  59. [59] Raghu Maithra and Schmidt Eric. 2020. A survey of deep learning for scientific discovery. arXiv preprint arXiv:2003.11755 (2020).Google ScholarGoogle Scholar
  60. [60] Schmidhuber Jürgen. 2015. Deep learning in neural networks: An overview. Neural Networks 61 (2015), 85117.Google ScholarGoogle ScholarDigital LibraryDigital Library
  61. [61] Zhang Qingchen, Yang Laurence T., et al. 2018. A survey on deep learning for big data. Information Fusion 42 (2018), 146157.Google ScholarGoogle ScholarCross RefCross Ref
  62. [62] Butepage Judith, Black Michael J., et al. 2017. Deep representation learning for human motion prediction and classification. In Conference on Computer Vision and Pattern Recognition. 61586166.Google ScholarGoogle ScholarCross RefCross Ref
  63. [63] Lv Zhibin, Cui Feifei, et al. 2021. Anticancer peptides prediction with deep representation learning features. Briefings in Bioinformatics 22, 5 (2021), bbab008.Google ScholarGoogle Scholar
  64. [64] Tian Haiman, Tao Yudong, et al. 2019. Multimodal deep representation learning for video classification. International World Wide Web Conference 22, 3 (2019), 13251341.Google ScholarGoogle ScholarDigital LibraryDigital Library
  65. [65] Guo Yang, Wu Zhenyu, and Ji Yang. 2017. A hybrid deep representation learning model for time series classification and prediction. In International Conference on Big Data Computing and Communications. IEEE, 226231.Google ScholarGoogle ScholarCross RefCross Ref
  66. [66] You Yuning, Chen Tianlong, et al. 2020. When does self-supervision help graph convolutional networks?. In International Conference on Machine Learning. PMLR.Google ScholarGoogle Scholar
  67. [67] Ericsson Linus, Gouk Henry, et al. 2022. Self-supervised representation learning: Introduction, advances, and challenges. IEEE Signal Processing Magazine 39, 3 (2022), 4262.Google ScholarGoogle ScholarCross RefCross Ref
  68. [68] Zhong Guoqiang, Wang Li-Na, et al. 2016. An overview on data representation learning: From traditional feature learning to recent deep learning. The Journal of Finance and Data Science 2, 4 (2016), 265278.Google ScholarGoogle ScholarCross RefCross Ref
  69. [69] Bengio Yoshua, Courville Aaron, and Vincent Pascal. 2013. Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence (2013).Google ScholarGoogle ScholarDigital LibraryDigital Library
  70. [70] Zhang Daokun, Yin Jie, et al. 2018. Network representation learning: A survey. IEEE Transactions on Big Data 6, 1 (2018), 328.Google ScholarGoogle ScholarCross RefCross Ref
  71. [71] Li Bentian and Pi Dechang. 2020. Network representation learning: A systematic literature review. Neural Computing and Applications 32, 21 (2020), 1664716679.Google ScholarGoogle ScholarDigital LibraryDigital Library
  72. [72] Chen Fenxiao, Wang Yun-Cheng, et al. 2020. Graph representation learning: A survey. APSIPA Transactions on Signal and Information Processing 9 (2020).Google ScholarGoogle ScholarCross RefCross Ref
  73. [73] Ridgeway Karl. 2016. A survey of inductive biases for factorial representation-learning. CoRR abs/1612.05299 (2016). arXiv:1612.05299 http://arxiv.org/abs/1612.05299Google ScholarGoogle Scholar
  74. [74] Zhang Zhao, Zhang Yan, et al. 2021. A survey on concept factorization: From shallow to deep representation learning. Information Processing & Management 58, 3 (2021), 102534.Google ScholarGoogle ScholarDigital LibraryDigital Library
  75. [75] Oussidi Achraf and Elhassouny Azeddine. 2018. Deep generative models: Survey. In International Conference on Intelligent Systems and Computer Vision. IEEE, 18.Google ScholarGoogle ScholarCross RefCross Ref
  76. [76] Regenwetter Lyle, Nobari Amin Heyrani, and Ahmed Faez. 2022. Deep generative models in engineering design: A review. Journal of Mechanical Design 144, 7 (2022), 071704.Google ScholarGoogle ScholarCross RefCross Ref
  77. [77] Eigenschink Peter, Vamosi Stefan, et al. 2021. Deep generative models for synthetic data. Comput. Surveys (2021).Google ScholarGoogle Scholar
  78. [78] Harshvardhan G. M., Gourisaria Mahendra Kumar, et al. 2020. A comprehensive survey and analysis of generative models in machine learning. Computer Science Review 38 (2020), 100285.Google ScholarGoogle ScholarDigital LibraryDigital Library
  79. [79] Guo Xiaojie and Zhao Liang. 2020. A systematic survey on deep generative models for graph generation. arXiv preprint arXiv:2007.06686 (2020).Google ScholarGoogle Scholar
  80. [80] Zhu Yanqiao, Du Yuanqi, et al. 2022. A survey on deep graph generation: Methods and applications. arXiv preprint arXiv:2203.06714 (2022).Google ScholarGoogle Scholar
  81. [81] Faez Faezeh, Ommi Yassaman, et al. 2021. Deep graph generators: A survey. IEEE Access 9 (2021), 106675106702.Google ScholarGoogle ScholarCross RefCross Ref
  82. [82] Lu Zhihe, Li Zhihang, et al. 2017. Recent progress of face image synthesis. In Asian Conference on Pattern Recognition. IEEE, 712.Google ScholarGoogle ScholarCross RefCross Ref
  83. [83] Luo Sanbi. 2021. A survey on multimodal deep learning for image synthesis: Applications, methods, datasets, evaluation metrics, and results comparison. In International Conference on Innovation in Artificial Intelligence. 108120.Google ScholarGoogle ScholarDigital LibraryDigital Library
  84. [84] He Xiaodong and Deng Li. 2017. Deep learning for image-to-text generation: A technical overview. IEEE Signal Processing Magazine 34, 6 (2017), 109116.Google ScholarGoogle ScholarCross RefCross Ref
  85. [85] Rosa Gustavo H. de and Papa Joao P.. 2021. A survey on text generation using generative adversarial networks. Pattern Recognition (2021).Google ScholarGoogle Scholar
  86. [86] Cho Yin-Ping, Yang Fu-Rong, et al. 2021. A survey on recent deep learning-driven singing voice synthesis systems. In International Conference on Artificial Intelligence and Virtual Reality. IEEE, 319323.Google ScholarGoogle ScholarCross RefCross Ref
  87. [87] Tan Xu, Qin Tao, et al. 2021. A survey on neural speech synthesis. arXiv preprint arXiv:2106.15561 (2021).Google ScholarGoogle Scholar
  88. [88] Mu Zhaoxi, Yang Xinyu, and Dong Yizhuo. 2021. Review of end-to-end speech synthesis technology based on deep learning. arXiv preprint arXiv:2104.09995 (2021).Google ScholarGoogle Scholar
  89. [89] Pederson D.. 1984. A historical review of circuit simulation. IEEE Transactions on Circuits and Systems 31, 1 (1984), 103111.Google ScholarGoogle ScholarCross RefCross Ref
  90. [90] Guihaire Valérie and Hao Jin-Kao. 2008. Transit network design and scheduling: A global review. Transportation Research Part A: Policy and Practice 42, 10 (2008), 12511273.Google ScholarGoogle ScholarCross RefCross Ref
  91. [91] Toshevska Martina and Gievska Sonja. 2021. A review of text style transfer using deep learning. TAI (2021), 669–684.Google ScholarGoogle Scholar
  92. [92] Jin Di, Jin Zhijing, et al. 2022. Deep learning for text style transfer: A survey. Computational Linguistics 48, 1 (2022), 155205.Google ScholarGoogle ScholarCross RefCross Ref
  93. [93] Du Yuanqi, Fu Tianfan, et al. 2022. MolGenSurvey: A systematic survey in machine learning models for molecule design. arXiv preprint arXiv:2203.14500 (2022).Google ScholarGoogle Scholar
  94. [94] Kim Jintae, Park Sera, et al. 2021. Comprehensive survey of recent drug discovery using deep learning. International Journal of Molecular Sciences 22, 18 (2021), 9983.Google ScholarGoogle ScholarCross RefCross Ref
  95. [95] Kell Douglas B., Samanta Soumitra, and Swainston Neil. 2020. Deep learning and generative methods in cheminformatics and chemical biology: Navigating small molecule space intelligently. Biochemical Journal 477, 23 (2020), 45594580.Google ScholarGoogle ScholarCross RefCross Ref
  96. [96] Gao Wenhao, Fu Tianfan, et al. 2022. Sample efficiency matters: A benchmark for practical molecular optimization. arXiv preprint arXiv:2206.12411 (2022).Google ScholarGoogle Scholar
  97. [97] Box K. J. and Comer J. E. A.. 2008. Using measured pKa, LogP and solubility to investigate supersaturation and predict BCS class. Current Drug Metabolism 9, 9 (2008), 869878.Google ScholarGoogle ScholarCross RefCross Ref
  98. [98] Guo Xiaojie, Wang Shiyu, and Zhao Liang. 2022. Graph neural networks: Graph transformation. In Graph Neural Networks: Foundations, Frontiers, and Applications. Springer, 251275.Google ScholarGoogle ScholarCross RefCross Ref
  99. [99] Kingma Durk P., Mohamed Shakir, et al. 2014. Semi-supervised learning with deep generative models. In Conference on Neural Information Processing Systems, Ghahramani Z., Welling M., Cortes C., Lawrence N., and Weinberger K. Q. (Eds.), Vol. 27. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2014/file/d523773c6b194f37b938d340d5d02232-Paper.pdfGoogle ScholarGoogle Scholar
  100. [100] Cheung Brian, Livezey Jesse A., et al. 2014. Discovering hidden factors of variation in deep networks. arXiv preprint arXiv:1412.6583 (2014).Google ScholarGoogle Scholar
  101. [101] Odena Augustus, Olah Christopher, and Shlens Jonathon. 2017. Conditional image synthesis with auxiliary classifier gans. In International Conference on Machine Learning. PMLR, 26422651.Google ScholarGoogle ScholarDigital LibraryDigital Library
  102. [102] Ficler Jessica and Goldberg Yoav. 2017. Controlling linguistic style aspects in neural language generation. In Proceedings of the Workshop on Stylistic Variation. 94104.Google ScholarGoogle ScholarCross RefCross Ref
  103. [103] Esteban Cristóbal, Hyland Stephanie L., and Rätsch Gunnar. 2017. Real-valued (medical) time series generation with recurrent conditional GANs. arXiv preprint arXiv:1706.02633 (2017).Google ScholarGoogle Scholar
  104. [104] Bodla Navaneeth, Hua Gang, and Chellappa Rama. 2018. Semi-supervised FusedGAN for conditional image generation. In European Conference on Computer Vision.Google ScholarGoogle ScholarDigital LibraryDigital Library
  105. [105] Hoogeboom Emiel, Satorras Vıctor Garcia, Vignac Clément, and Welling Max. 2022. Equivariant diffusion for molecule generation in 3D. In International Conference on Machine Learning. PMLR, 88678887.Google ScholarGoogle Scholar
  106. [106] Nichol Alex, Dhariwal Prafulla, Ramesh Aditya, Shyam Pranav, Mishkin Pamela, McGrew Bob, Sutskever Ilya, and Chen Mark. 2021. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741 (2021).Google ScholarGoogle Scholar
  107. [107] Henter Gustav Eje, Lorenzo-Trueba Jaime, et al. 2018. Deep encoder-decoder models for unsupervised learning of controllable speech synthesis. arXiv preprint arXiv:1807.11470 (2018).Google ScholarGoogle Scholar
  108. [108] Tevet Guy, Raab Sigal, Gordon Brian, Shafir Yonatan, Cohen-Or Daniel, and Bermano Amit H.. 2022. Human motion diffusion model. arXiv preprint arXiv:2209.14916 (2022).Google ScholarGoogle Scholar
  109. [109] Rothchild Daniel, Tamkin Alex, Yu Julie, Misra Ujval, and Gonzalez Joseph. 2021. C5T5: Controllable generation of organic molecules with transformers. arXiv preprint arXiv:2108.10307 (2021).Google ScholarGoogle Scholar
  110. [110] Li Yujia, Vinyals Oriol, et al. 2018. Learning deep generative models of graphs. arXiv preprint arXiv:1803.03324 (2018).Google ScholarGoogle Scholar
  111. [111] Yang Ximing, Wu Yuan, et al. 2021. CPCGAN: A controllable 3D point cloud generative adversarial network with semantic label generating. In AAAI Conference on Artificial Intelligence, Vol. 35. 31543162.Google ScholarGoogle ScholarCross RefCross Ref
  112. [112] Xu Lei, Skoularidou Maria, et al. 2019. Modeling tabular data using conditional GAN. Conference on Neural Information Processing Systems 32 (2019).Google ScholarGoogle Scholar
  113. [113] Keskar Nitish Shirish, McCann Bryan, et al. 2019. CTRL: A conditional transformer language model for controllable generation. arXiv preprint arXiv:1909.05858 (2019).Google ScholarGoogle Scholar
  114. [114] Prabhumoye Shrimai, Black Alan W., and Salakhutdinov Ruslan. 2020. Exploring controllable text generation techniques. In International Conference on Computational Linguistics. COLING, Barcelona, Spain (Online), 114. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  115. [115] Xu Peng, Patwary Mostofa, Shoeybi Mohammad, Puri Raul, Fung Pascale, Anandkumar Anima, and Catanzaro Bryan. 2020. MEGATRON-CNTRL: Controllable story generation with external knowledge using large-scale language models. arXiv preprint arXiv:2010.00840 (2020).Google ScholarGoogle Scholar
  116. [116] Jin Wengong, Barzilay Regina, and Jaakkola Tommi. 2020. Multi-objective molecule generation using interpretable substructures. In International Conference on Machine Learning. PMLR, 48494859.Google ScholarGoogle Scholar
  117. [117] Mokhayeri Fania, Kamali Kaveh, and Granger Eric. 2020. Cross-Domain face synthesis using a controllable GAN. In Winter Conference on Applications of Computer Vision. 252260.Google ScholarGoogle ScholarCross RefCross Ref
  118. [118] Huang Rongjie, Lam Max W. Y., et al. 2022. FastDiff: A fast conditional diffusion model for high-quality speech synthesis. In International Joint Conferences on Artificial Intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  119. [119] Sohn Kihyuk, Lee Honglak, and Yan Xinchen. 2015. Learning structured output representation using deep conditional generative models. In Conference on Neural Information Processing Systems, Cortes C., Lawrence N., Lee D., Sugiyama M., and Garnett R. (Eds.), Vol. 28. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2015/file/8d55a249e6baa5c06772297520da2051-Paper.pdfGoogle ScholarGoogle Scholar
  120. [120] Shao Huajie, Wang Jun, et al. 2021. Controllable and diverse text generation in e-commerce. In International World Wide Web Conference. 23922401.Google ScholarGoogle ScholarDigital LibraryDigital Library
  121. [121] Vasquez Sean and Lewis Mike. 2019. MelNet: A generative model for audio in the frequency domain. arXiv preprint arXiv:1906.01083 (2019).Google ScholarGoogle Scholar
  122. [122] Zang Chengxi and Wang Fei. 2020. MoFlow: An invertible flow model for generating molecular graphs. In Special Interest Group on Knowledge Discovery and Data Mining. 617626.Google ScholarGoogle ScholarDigital LibraryDigital Library
  123. [123] Moon Jaeuk, Jung Seungwon, et al. 2020. Conditional tabular GAN-based two-stage data generation scheme for short-term load forecasting. IEEE Access 8 (2020), 205327205339.Google ScholarGoogle ScholarCross RefCross Ref
  124. [124] Chen Ke, Wang Cheng-i, et al. 2020. Music SketchNet: Controllable music generation via factorized representations of pitch and rhythm. In International Society for Music Information Retrieval.Google ScholarGoogle Scholar
  125. [125] Shoshan Alon, Bhonker Nadav, et al. 2021. GAN-control: Explicitly controllable GANs. In International Conference on Computer Vision. 1408314093.Google ScholarGoogle ScholarCross RefCross Ref
  126. [126] Lu You and Huang Bert. 2020. Structured output learning with conditional generative flows. In AAAI Conference on Artificial Intelligence, Vol. 34. 50055012.Google ScholarGoogle ScholarCross RefCross Ref
  127. [127] Ardizzone Lynton, Lüth Carsten, et al. 2019. Guided image generation with conditional invertible neural networks. arXiv preprint arXiv:1907.02392 (2019).Google ScholarGoogle Scholar
  128. [128] Gómez-Bombarelli Rafael, Wei Jennifer N., et al. 2018. Automatic chemical design using a data-driven continuous representation of molecules. ACS Central Science 4, 2 (2018), 268276.Google ScholarGoogle ScholarCross RefCross Ref
  129. [129] Choi Jooyoung, Kim Sungwon, et al. 2021. ILVR: Conditioning method for denoising diffusion probabilistic models. International Conference on Computer Vision (2021).Google ScholarGoogle Scholar
  130. [130] Kang Yanfei, Hyndman Rob J., et al. 2018. Efficient Generation of Time Series with Diverse and Controllable Characteristics. Technical Report. Monash University, Department of Econometrics and Business Statistics.Google ScholarGoogle Scholar
  131. [131] Liu Meng, Yan Keqiang, et al. 2021. GraphEBM: Molecular graph generation with energy-based models. arXiv preprint arXiv:2102.00546 (2021).Google ScholarGoogle Scholar
  132. [132] Deng Yu, Yang Jiaolong, et al. 2020. Disentangled and controllable face image generation via 3D imitative-contrastive learning. In Conference on Computer Vision and Pattern Recognition. 51545163.Google ScholarGoogle ScholarCross RefCross Ref
  133. [133] Park Noseong, Mohammadi Mahmoud, et al. 2018. Data synthesis based on generative adversarial networks. Proc. VLDB Endow. 11, 10 (Jun.2018), 10711083. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  134. [134] Mathieu Michael F., Zhao Junbo Jake, et al. 2016. Disentangling factors of variation in deep representation using adversarial training. In Conference on Neural Information Processing Systems, Lee D., Sugiyama M., Luxburg U., Guyon I., and Garnett R. (Eds.), Vol. 29. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2016/file/ef0917ea498b1665ad6c701057155abe-Paper.pdfGoogle ScholarGoogle Scholar
  135. [135] Johnson Matthew J., Duvenaud David K., et al. 2016. Composing graphical models with neural networks for structured representations and fast inference. In Conference on Neural Information Processing Systems, Lee D., Sugiyama M., Luxburg U., Guyon I., and Garnett R. (Eds.), Vol. 29. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2016/file/7d6044e95a16761171b130dcb476a43e-Paper.pdfGoogle ScholarGoogle Scholar
  136. [136] Upchurch Paul, Gardner Jacob, et al. 2017. Deep feature interpolation for image content changes. In Conference on Computer Vision and Pattern Recognition. 70647073.Google ScholarGoogle ScholarCross RefCross Ref
  137. [137] Kingma Durk P. and Dhariwal Prafulla. 2018. Glow: Generative Flow with Invertible 1x1 Convolutions. In Conference on Neural Information Processing Systems, Bengio S., Wallach H., et al. (Eds.), Vol. 31. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2018/file/d139db6a236200b21cc7f752979132d0-Paper.pdfGoogle ScholarGoogle Scholar
  138. [138] Camino Ramiro, Hammerschmidt Christian, and State Radu. 2018. Generating multi-categorical samples with generative adversarial networks. arXiv preprint arXiv:1807.01202 (2018).Google ScholarGoogle Scholar
  139. [139] Wang Ziyu, Wang Dingsu, et al. 2020. Learning interpretable representation for controllable polyphonic music generation. In International Society for Music Information Retrieval 2020, Cumming Julie, Lee Jin Ha, McFee Brian, Schedl Markus, Devaney Johanna, McKay Cory, Zangerle Eva, and Reuse Timothy de (Eds.). 662669. http://archives.ismir.net/ismir2020/paper/000094.pdfGoogle ScholarGoogle Scholar
  140. [140] Ma Tengfei, Chen Jie, and Xiao Cao. 2018. Constrained generation of semantically valid graphs via regularizing variational autoencoders. Conference on Neural Information Processing Systems 31 (2018).Google ScholarGoogle Scholar
  141. [141] Goetschalckx Lore, Andonian Alex, et al. 2019. GANalyze: Toward visual definitions of cognitive image properties. In International Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  142. [142] Möllenhoff Thomas and Cremers Daniel. 2019. Flat metric minimization with applications in generative modeling. In International Conference on Machine Learning. PMLR, 46264635.Google ScholarGoogle Scholar
  143. [143] Madhawa Kaushalya, Ishiguro Katushiko, et al. 2019. GraphNVP: An invertible flow model for generating molecular graphs. arXiv preprint arXiv:1905.11600 (2019).Google ScholarGoogle Scholar
  144. [144] Shen Yujun and Zhou Bolei. 2021. Closed-form factorization of latent semantics in GANs. In Conference on Computer Vision and Pattern Recognition. 15321540.Google ScholarGoogle ScholarCross RefCross Ref
  145. [145] Du Yuanqi, Liu Xian, et al. 2022. ChemSpacE: Toward steerable and interpretable chemical space exploration. In International Conference on Learning Representations 2022 Machine Learning for Drug Discovery.Google ScholarGoogle Scholar
  146. [146] Guo Xiaojie, Du Yuanqi, and Zhao Liang. 2021. Deep generative models for spatial networks. In Special Interest Group on Knowledge Discovery and Data Mining. 505515.Google ScholarGoogle ScholarDigital LibraryDigital Library
  147. [147] Du Yuanqi, Guo Xiaojie, et al. 2022. Disentangled spatiotemporal graph generative models. AAAI Conference on Artificial Intelligence 36, 6 (Jun.2022), 65416549. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  148. [148] Guo Xiaojie, Du Yuanqi, et al. 2021. Generating tertiary protein structures via interpretable graph variational autoencoders. Bioinformatics Advances 1, 1 (2021), vbab036.Google ScholarGoogle ScholarCross RefCross Ref
  149. [149] Pati Ashis and Lerch Alexander. 2021. Is disentanglement enough? On latent representations for controllable music generation. In International Society for Music Information Retrieval. Online.Google ScholarGoogle Scholar
  150. [150] Li Shidi, Liu Miaomiao, and Walder Christian. 2022. EditVAE: Unsupervised parts-aware controllable 3D point cloud shape generation. AAAI Conference on Artificial Intelligence 36, 2 (Jun.2022), 13861394. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  151. [151] Radford Alec, Metz Luke, and Chintala Soumith. 2016. Unsupervised representation learning with deep convolutional generative adversarial networks. In International Conference on Learning Representations, Bengio Yoshua and LeCun Yann (Eds.). http://arxiv.org/abs/1511.06434Google ScholarGoogle Scholar
  152. [152] Du Yuanqi, Guo Xiaojie, et al. 2020. Interpretable molecule generation via disentanglement learning. In Proceedings of the 11th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics. 18.Google ScholarGoogle ScholarDigital LibraryDigital Library
  153. [153] Zhou Hang, Liu Yu, et al. 2019. Talking face generation by adversarially disentangled audio-visual representation. In AAAI Conference on Artificial Intelligence, Vol. 33. 92999306.Google ScholarGoogle ScholarDigital LibraryDigital Library
  154. [154] Guo Xiaojie, Zhao Liang, et al. 2020. Interpretable deep graph generation with node-edge co-disentanglement. In Special Interest Group on Knowledge Discovery and Data Mining.Google ScholarGoogle ScholarDigital LibraryDigital Library
  155. [155] Du Yuanqi, Wang Yinkai, et al. 2021. Deep latent-variable models for controllable molecule generation. In International Conference on Bioinformatics and Biomedicine. IEEE, 372375.Google ScholarGoogle ScholarCross RefCross Ref
  156. [156] Du Yuanqi, Guo Xiaojie, et al. 2022. Small molecule generation via disentangled representation learning. Bioinformatics (Oxford, England) (2022), btac296.Google ScholarGoogle Scholar
  157. [157] Plumerault Antoine, Borgne Hervé Le, and Hudelot Céline. 2020. Controlling generative models with continuous factors of variations. In International Conference on Learning Representations. https://openreview.net/forum?id=H1laeJrKDBGoogle ScholarGoogle Scholar
  158. [158] Locatello Francesco, Tschannen Michael, et al. 2020. Disentangling factors of variations using few labels. In International Conference on Learning Representations. https://openreview.net/forum?id=SygagpEKwBGoogle ScholarGoogle Scholar
  159. [159] Jahanian Ali, Chai Lucy, and Isola Phillip. 2020. On the “steerability” of generative adversarial networks. In International Conference on Learning Representations. https://openreview.net/forum?id=HylsTT4FvBGoogle ScholarGoogle Scholar
  160. [160] Härkönen Erik, Hertzmann Aaron, et al. 2020. GANSpace: Discovering interpretable GAN controls. Conference on Neural Information Processing Systems 33 (2020), 98419850.Google ScholarGoogle Scholar
  161. [161] Yang Jie, Mo Kaichun, et al. 2020. DSM-Net: Disentangled structured mesh net for controllable generation of fine geometry. arXiv preprint arXiv:2008.05440 2, 3 (2020).Google ScholarGoogle Scholar
  162. [162] Xu Peng, Cheung Jackie Chi Kit, and Cao Yanshuai. 2020. On variational learning of controllable representations for text without supervision. In International Conference on Machine Learning. PMLR, 1053410543.Google ScholarGoogle Scholar
  163. [163] Tan Hao Hao and Herremans Dorien. 2020. Music FaderNets: Controllable music generation based on high-level features via low-level feature modelling. In International Society for Music Information Retrieval.Google ScholarGoogle Scholar
  164. [164] Parthasarathy Dhasarathy, Bäckstrom Karl, et al. 2020. Controlled time series generation for automotive software-in-the-loop testing using GANs. In IEEE International Conference on Artificial Intelligence Testing. IEEE, 3946.Google ScholarGoogle ScholarCross RefCross Ref
  165. [165] Wang Shiyu, Guo Xiaojie, and Zhao Liang. 2022. Deep generative model for periodic graphs. arXiv preprint arXiv:2201.11932 (2022).Google ScholarGoogle Scholar
  166. [166] Nigam AkshatKumar, Friederich Pascal, et al. 2020. Augmenting genetic algorithms with deep neural networks for exploring the chemical space. In International Conference on Learning Representations. https://openreview.net/forum?id=H1lmyRNFvrGoogle ScholarGoogle Scholar
  167. [167] Khalifa Muhammad, Elsahar Hady, and Dymetman Marc. 2021. A distributional approach to controlled text generation. In International Conference on Learning Representations. https://openreview.net/forum?id=jWkw45-9AbLGoogle ScholarGoogle Scholar
  168. [168] Fu Tianfan, Gao Wenhao, et al. 2022. Differentiable scaffolding tree for molecule optimization. In International Conference on Learning Representations. https://openreview.net/forum?id=w_drCosT76Google ScholarGoogle Scholar
  169. [169] Kang Yanfei, Hyndman Rob J., and Li Feng. 2020. GRATIS: GeneRAting time series with diverse and controllable characteristics. Statistical Analysis and Data Mining: The ASA Data Science Journal 13, 4 (2020), 354376.Google ScholarGoogle ScholarDigital LibraryDigital Library
  170. [170] Dathathri Sumanth, Madotto Andrea, et al. 2020. Plug and play language models: A simple approach to controlled text generation. In International Conference on Learning Representations. https://openreview.net/forum?id=H1edEyBKDSGoogle ScholarGoogle Scholar
  171. [171] Yu Lantao, Zhang Weinan, et al. 2017. SeqGAN: Sequence generative adversarial nets with policy gradient. In AAAI Conference on Artificial Intelligence, Vol. 31.Google ScholarGoogle ScholarCross RefCross Ref
  172. [172] Hu Zhiting, Yang Zichao, et al. 2018. Deep generative models with learnable knowledge constraints. Conference on Neural Information Processing Systems 31 (2018).Google ScholarGoogle Scholar
  173. [173] Putin Evgeny, Asadulaev Arip, et al. 2018. Reinforced adversarial neural computer for de novo molecular design. Journal of Chemical Information and Modeling 58, 6 (2018), 11941204.Google ScholarGoogle ScholarCross RefCross Ref
  174. [174] Tambwekar Pradyumna, Dhuliawala Murtaza, et al. 2019. Controllable neural story plot generation via reinforcement learning. In International Joint Conferences on Artificial Intelligence.Google ScholarGoogle Scholar
  175. [175] Shi Chence, Xu Minkai, et al. 2020. GraphAF: A flow-based autoregressive model for molecular graph generation. In International Conference on Learning Representations. https://openreview.net/forum?id=S1esMkHYPrGoogle ScholarGoogle Scholar
  176. [176] Yang Xiufeng, Aasawat Tanuj, and Yoshizoe Kazuki. 2021. Practical massively parallel Monte-Carlo tree search applied to molecular design. In International Conference on Learning Representations. https://openreview.net/forum?id=6k7VdojAIKGoogle ScholarGoogle Scholar
  177. [177] Samanta Bidisha, De Abir, et al. 2020. NeVAE: A deep generative model for molecular graphs. Journal of Machine Learning Research (2020).Google ScholarGoogle Scholar
  178. [178] Wang Jike, Hsieh Chang-Yu, et al. 2021. Multi-constraint molecular generation based on conditional transformer, knowledge distillation and reinforcement learning. Nature Machine Intelligence 3, 10 (2021), 914922.Google ScholarGoogle ScholarCross RefCross Ref
  179. [179] Gebauer Niklas, Gastegger Michael, and Schütt Kristof. 2019. Symmetry-adapted generation of 3D point sets for the targeted discovery of molecules. Conference on Neural Information Processing Systems 32 (2019).Google ScholarGoogle Scholar
  180. [180] Segler Marwin H. S., Kogej Thierry, et al. 2018. Generating focused molecule libraries for drug discovery with recurrent neural networks. ACS Central Science 4, 1 (2018), 120131.Google ScholarGoogle ScholarCross RefCross Ref
  181. [181] Prabhumoye Shrimai, Tsvetkov Yulia, et al. 2018. Style transfer through back-translation. In Proc. ACL.Google ScholarGoogle ScholarCross RefCross Ref
  182. [182] Jin Wengong, Yang Kevin, et al. 2019. Learning multimodal graph-to-graph translation for molecule optimization. In International Conference on Learning Representations. https://openreview.net/forum?id=B1xJAsA5F7Google ScholarGoogle Scholar
  183. [183] Sood Rewa, Topiwala Binit, et al. 2018. An application of generative adversarial networks for super resolution medical imaging. In International Conference on Machine Learning and Applications. IEEE, 326331.Google ScholarGoogle ScholarCross RefCross Ref
  184. [184] Luo Youzhi, Yan Keqiang, and Ji Shuiwang. 2021. GraphDF: A discrete flow model for molecular graph generation. In International Conference on Machine Learning. PMLR, 71927203.Google ScholarGoogle Scholar
  185. [185] Fu Tianfan, Xiao Cao, and Sun Jimeng. 2020. Core: Automatic molecule optimization using copy & refine strategy. In AAAI Conference on Artificial Intelligence, Vol. 34. 638645.Google ScholarGoogle ScholarCross RefCross Ref
  186. [186] Sudhakar Akhilesh, Upadhyay Bhargav, and Maheswaran Arjun. 2019. Transforming delete, retrieve, generate approach for controlled text style transfer. In Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle ScholarCross RefCross Ref
  187. [187] Hukkelås Håkon, Mester Rudolf, and Lindseth Frank. 2019. DeepPrivacy: A generative adversarial network for face anonymization. In International symposium on visual computing. Springer, 565578.Google ScholarGoogle ScholarDigital LibraryDigital Library
  188. [188] Zhou Zhenpeng, Li Xiaocheng, and Zare Richard N.. 2017. Optimizing chemical reactions with deep reinforcement learning. ACS Central Science 3, 12 (2017), 13371344.Google ScholarGoogle ScholarCross RefCross Ref
  189. [189] Tits Noé, Haddad Kevin El, and Dutoit Thierry. 2019. Exploring transfer learning for low resource emotional TTS. In Proceedings of SAI Intelligent Systems Conference. Springer, 5260.Google ScholarGoogle Scholar
  190. [190] Liu Rui, Sisman Berrak, and Li Haizhou. 2021. Reinforcement learning for emotional text-to-speech synthesis with improved emotion discriminability. In Interspeech.Google ScholarGoogle Scholar
  191. [191] Hsu Wei-Ning, Zhang Yu, et al. 2019. Hierarchical generative modeling for controllable speech synthesis. In International Conference on Learning Representations. https://openreview.net/forum?id=rygkk305YQGoogle ScholarGoogle Scholar
  192. [192] Chen Mingda, Tang Qingming, et al. 2019. Controllable paraphrase generation with a syntactic exemplar. In Proc. of ACL.Google ScholarGoogle ScholarCross RefCross Ref
  193. [193] Liang Dong, Wang Rui, et al. 2019. PCGAN: Partition-controlled human image generation. In AAAI Conference on Artificial Intelligence, Vol. 33. 86988705.Google ScholarGoogle ScholarDigital LibraryDigital Library
  194. [194] Chen Lele, Maddox Ross K., et al. 2019. Hierarchical cross-modal talking face generation with dynamic pixel-wise loss. In Conference on Computer Vision and Pattern Recognition. 78327841.Google ScholarGoogle ScholarCross RefCross Ref
  195. [195] Maximov Maxim, Elezi Ismail, and Leal-Taixé Laura. 2020. CIAGAN: Conditional identity anonymization generative adversarial networks. In Conference on Computer Vision and Pattern Recognition. 54475456.Google ScholarGoogle ScholarCross RefCross Ref
  196. [196] Li Tao, Yang Shan, et al. 2021. Controllable emotion transfer for end-to-end speech synthesis. In International Symposium on Chinese Spoken Language Processing. IEEE, 15.Google ScholarGoogle ScholarCross RefCross Ref
  197. [197] Huang Kuan-Hao and Chang Kai-Wei. 2021. Generating syntactically controlled paraphrases without using annotated parallel pairs. In Conference of the European Chapter of the Association for Computational Linguistics.Google ScholarGoogle ScholarCross RefCross Ref
  198. [198] Pumarola Albert, Popov Stefan, et al. 2020. C-flow: Conditional generative flow models for images and 3D point clouds. In Conference on Computer Vision and Pattern Recognition. 79497958.Google ScholarGoogle ScholarCross RefCross Ref
  199. [199] Yang Zichao, Hu Zhiting, et al. 2018. Unsupervised text style transfer using language models as discriminators. Conference on Neural Information Processing Systems 31 (2018).Google ScholarGoogle Scholar
  200. [200] Valle Rafael, Li Jason, et al. 2020. Mellotron: Multispeaker expressive voice synthesis by conditioning on rhythm, pitch and global style tokens. In IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 61896193.Google ScholarGoogle ScholarCross RefCross Ref
  201. [201] Kurihara Kiyoshi, Seiyama Nobumasa, and Kumano Tadashi. 2021. Prosodic features control by symbols as input of sequence-to-sequence acoustic modeling for neural TTS. IEICE Transactions on Information and Systems 104, 2 (2021), 302311.Google ScholarGoogle ScholarCross RefCross Ref
  202. [202] Kim Minchan, Cheon Sung Jun, et al. 2021. Expressive text-to-speech using style tag. In Interspeech.Google ScholarGoogle Scholar
  203. [203] Bian Yanyao, Chen Changbin, et al. 2019. Multi-reference tacotron by intercross training for style disentangling, transfer and control in speech synthesis. arXiv preprint arXiv:1904.02373 (2019).Google ScholarGoogle Scholar
  204. [204] Inoue Katsuki, Hara Sunao, et al. 2021. Model architectures to extrapolate emotional expressions in DNN-based text-to-speech. Speech Communication 126 (2021), 3543.Google ScholarGoogle ScholarCross RefCross Ref
  205. [205] Cai Xiong, Dai Dongyang, et al. 2021. Emotion controllable speech synthesis using emotion-unlabeled dataset with the assistance of cross-domain speech emotion recognition. In IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 57345738.Google ScholarGoogle ScholarCross RefCross Ref
  206. [206] Krause Ben, Gotmare Akhilesh Deepak, et al. 2020. GeDi: Generative discriminator guided sequence generation. In Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle Scholar
  207. [207] Song Yang, Zhu Jingwen, et al. 2019. Talking face generation by conditional recurrent adversarial network. In International Joint Conferences on Artificial Intelligence. 919925.Google ScholarGoogle ScholarCross RefCross Ref
  208. [208] Kwon Ohsung, Jang Inseon, et al. 2019. An effective style token weight control technique for end-to-end emotional speech synthesis. IEEE Signal Processing Letters 26, 9 (2019), 13831387.Google ScholarGoogle ScholarCross RefCross Ref
  209. [209] Tits Noé. 2019. A methodology for controlling the emotional expressiveness in synthetic speech-a deep learning approach. In International Conference on Affective Computing and Intelligent Interaction Workshops and Demos. IEEE, 15.Google ScholarGoogle ScholarCross RefCross Ref
  210. [210] Sini Aghilas, Maguer Sébastien Le, et al. 2020. Introducing prosodic speaker identity for a better expressive speech synthesis control. In 10th International Conference on Speech Prosody 2020. ISCA, 935939.Google ScholarGoogle ScholarCross RefCross Ref
  211. [211] Li Juncen, Jia Robin, et al. 2018. Delete, retrieve, generate: A simple approach to sentiment and style transfer. In Annual Conference of the North American Chapter of the Association for Computational Linguistics.Google ScholarGoogle ScholarCross RefCross Ref
  212. [212] Fabbro Giorgio, Golkov Vladimir, et al. 2020. Speech synthesis and control using differentiable DSP. arXiv preprint arXiv:2010.15084 (2020).Google ScholarGoogle Scholar
  213. [213] John Vineet, Mou Lili, Bahuleyan Hareesh, and Vechtomova Olga. 2019. Disentangled representation learning for non-parallel text style transfer. In Conference of the European Chapter of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 424434. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  214. [214] Abdal Rameen, Zhu Peihao, et al. 2021. StyleFlow: Attribute-conditioned exploration of StyleGAN-generated images using conditional continuous normalizing flows. TOG 40, 3 (2021), 121.Google ScholarGoogle ScholarDigital LibraryDigital Library
  215. [215] Zhu Jiapeng, Feng Ruili, et al. 2021. Low-rank subspaces in GANs. Conference on Neural Information Processing Systems 34 (2021).Google ScholarGoogle Scholar
  216. [216] Xia Weihao, Zhang Yulun, et al. 2022. GAN inversion: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence01 (Jun.2022), 117. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  217. [217] Thermos Spyridon, Liu Xiao, et al. 2021. Controllable cardiac synthesis via disentangled anatomy arithmetic. In International Conference on Medical Image Computing and Computer Assisted Intervention. Springer, 160170.Google ScholarGoogle ScholarDigital LibraryDigital Library
  218. [218] Chang Haw-Shiuan, Yuan Jiaming, Iyyer Mohit, and McCallum Andrew. 2021. Changing the mind of transformers for topically-controllable language generation. In Conference of the European Chapter of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 26012611. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  219. [219] Shen Yujun, Gu Jinjin, et al. 2020. Interpreting the latent space of GANs for semantic face editing. In Conference on Computer Vision and Pattern Recognition. 92439252.Google ScholarGoogle ScholarCross RefCross Ref
  220. [220] Qiao Lin, Yan Jianhao, et al. 2020. A sentiment-controllable topic-to-essay generator with topic knowledge graph. In Findings of Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle ScholarCross RefCross Ref
  221. [221] Guu Kelvin, Hashimoto Tatsunori B., et al. 2018. Generating sentences by editing prototypes. Transactions of the Association for Computational Linguistics 6 (2018), 437450.Google ScholarGoogle ScholarCross RefCross Ref
  222. [222] Ren Yi, Ruan Yangjun, et al. 2019. FastSpeech: Fast, robust and controllable text to speech. Conference on Neural Information Processing Systems 32 (2019).Google ScholarGoogle Scholar
  223. [223] Luo Xuan, Takamichi Shinnosuke, et al. 2021. Controllable text-to-speech synthesis using prosodic features and emotion soft-label. Sython.org.Google ScholarGoogle Scholar
  224. [224] Raitio Tuomo, Rasipuram Ramya, and Castellani Dan. 2020. Controllable neural text-to-speech synthesis using intuitive prosodic features. In INTERSPEECH.Google ScholarGoogle Scholar
  225. [225] Cui Chenye, Ren Yi, et al. 2021. EMOVIE: A Mandarin emotion speech dataset with a simple emotional text-to-speech model. In Proc. Interspeech 2021. 27662770. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  226. [226] Engel Jesse, Hantrakul Lamtharn (Hanoi), et al. 2020. DDSP: Differentiable digital signal processing. In International Conference on Learning Representations. https://openreview.net/forum?id=B1x1ma4tDrGoogle ScholarGoogle Scholar
  227. [227] Sutherland Danica J., Tung Hsiao-Yu, et al. 2017. Generative models and model criticism via optimized maximum mean discrepancy. In International Conference on Learning Representations. https://openreview.net/forum?id=HJWHIKqglGoogle ScholarGoogle Scholar
  228. [228] Zhang Liming, Zhao Liang, et al. 2021. TG-GAN: Continuous-time temporal graph deep generative models with time-validity constraints. In International World Wide Web Conference. 21042116.Google ScholarGoogle ScholarDigital LibraryDigital Library
  229. [229] Song Yang and Ermon Stefano. 2019. Generative modeling by estimating gradients of the data distribution. Conference on Neural Information Processing Systems 32 (2019).Google ScholarGoogle Scholar
  230. [230] Cheng Yu, Gong Yongshun, et al. 2021. Molecular design in drug discovery: A comprehensive review of deep generative models. Briefings in Bioinformatics 22, 6 (2021), bbab344.Google ScholarGoogle ScholarCross RefCross Ref
  231. [231] Cui Hejie, Lu Jiaying, et al. 2022. How can graph neural networks help document retrieval: A case study on cord19 with concept map generation. In European Conference on Information Retrieval. Springer, 7583.Google ScholarGoogle ScholarDigital LibraryDigital Library
  232. [232] Yacouby Reda and Axman Dustin. 2020. Probabilistic extension of precision, recall, and F1 score for more thorough evaluation of classification models. In Proceedings of the First Workshop on Evaluation and Comparison of NLP Systems. 7991.Google ScholarGoogle ScholarCross RefCross Ref
  233. [233] Kedzie Chris and McKeown Kathleen. 2020. Controllable meaning representation to text generation: Linearization and data augmentation strategies. In Conference on Empirical Methods in Natural Language Processing. 51605185.Google ScholarGoogle ScholarCross RefCross Ref
  234. [234] Rosenberg Andrew and Ramabhadran Bhuvana. 2017. Bias and statistical significance in evaluating speech synthesis with mean opinion scores. In Interspeech. 39763980.Google ScholarGoogle ScholarCross RefCross Ref
  235. [235] Ho Jonathan and Salimans Tim. 2022. Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598 (2022).Google ScholarGoogle Scholar
  236. [236] Griffiths Ryan-Rhys and Hernández-Lobato José Miguel. 2020. Constrained Bayesian optimization for automatic chemical design using variational autoencoders. Chemical Science 11, 2 (2020), 577586.Google ScholarGoogle ScholarCross RefCross Ref
  237. [237] Song Yang, Sohl-Dickstein Jascha, Kingma Diederik P., Kumar Abhishek, Ermon Stefano, and Poole Ben. 2020. Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456 (2020).Google ScholarGoogle Scholar
  238. [238] Dhariwal Prafulla and Nichol Alexander. 2021. Diffusion models beat GANs on image synthesis. Advances in Neural Information Processing Systems 34 (2021), 87808794.Google ScholarGoogle Scholar
  239. [239] Kim Heeseung, Kim Sungwon, and Yoon Sungroh. 2022. Guided-TTS: A diffusion model for text-to-speech via classifier guidance. In International Conference on Machine Learning. PMLR, 1111911133.Google ScholarGoogle Scholar
  240. [240] Kawar Bahjat, Ganz Roy, and Elad Michael. 2022. Enhancing diffusion-based image synthesis with robust classifier guidance. arXiv preprint arXiv:2208.08664 (2022).Google ScholarGoogle Scholar
  241. [241] Zhuang Fuzhen, Qi Zhiyuan, et al. 2020. A comprehensive survey on transfer learning. Proc. IEEE 109, 1 (2020), 4376.Google ScholarGoogle ScholarCross RefCross Ref
  242. [242] Arulkumaran Kai, Deisenroth Marc Peter, et al. 2017. Deep reinforcement learning: A brief survey. IEEE Signal Processing Magazine 34, 6 (2017), 2638.Google ScholarGoogle ScholarCross RefCross Ref
  243. [243] Ganchev Kuzman, Graça Joao, et al. 2010. Posterior regularization for structured latent variable models. Journal of Machine Learning Research 11 (2010), 20012049.Google ScholarGoogle ScholarDigital LibraryDigital Library
  244. [244] Sanchez-Lengeling Benjamin, Outeiral Carlos, et al. 2017. Optimizing distributions over molecular space. An objective-reinforced generative adversarial network for inverse-design chemistry (ORGANIC). Chemrxiv.org.Google ScholarGoogle Scholar
  245. [245] Zhang Zhu, Ma Jianxin, et al. 2021. UFC-BERT: Unifying multi-modal controls for conditional image synthesis. Conference on Neural Information Processing Systems 34 (2021).Google ScholarGoogle Scholar
  246. [246] Anderson Eric, Veith Gilman D., and Weininger David. 1987. SMILES, a Line Notation and Computerized Interpreter for Chemical Structures. US Environmental Protection Agency, Environmental Research Laboratory.Google ScholarGoogle Scholar
  247. [247] Walters W. Patrick and Barzilay Regina. 2020. Applications of deep learning in molecule generation and molecular property prediction. Accounts of Chemical Research 54, 2 (2020), 263270.Google ScholarGoogle ScholarCross RefCross Ref
  248. [248] Elton Daniel C., Boukouvalas Zois, et al. 2019. Deep learning for molecular design—a review of the state of the art. Molecular Systems Design & Engineering 4, 4 (2019), 828849.Google ScholarGoogle ScholarCross RefCross Ref
  249. [249] Meyers Joshua, Fabian Benedek, and Brown Nathan. 2021. De novo molecular design and generative models. Drug Discovery Today 26, 11 (2021), 27072715.Google ScholarGoogle ScholarCross RefCross Ref
  250. [250] Griffiths Ryan-Rhys, Klarner Leo, et al. GAUCHE: A library for Gaussian processes in chemistry. In International Conference on Machine Learning 2022 2nd AI for Science Workshop.Google ScholarGoogle Scholar
  251. [251] Xiong Zhaoping, Wang Dingyan, et al. 2019. Pushing the boundaries of molecular representation for drug discovery with the graph attention mechanism. Journal of Medicinal Chemistry 63, 16 (2019), 87498760.Google ScholarGoogle ScholarCross RefCross Ref
  252. [252] Shen Wan Xiang, Zeng Xian, et al. 2021. Out-of-the-box deep learning prediction of pharmaceutical properties by broadly learned knowledge-based molecular representations. Nature Machine Intelligence 3, 4 (2021), 334343.Google ScholarGoogle ScholarCross RefCross Ref
  253. [253] Simonovsky Martin and Komodakis Nikos. 2018. GraphVAE: Towards generation of small graphs using variational autoencoders. In International Conference on Artificial Neural Networks. Springer, 412422.Google ScholarGoogle ScholarCross RefCross Ref
  254. [254] You Jiaxuan, Ying Rex, et al. 2018. GraphRNN: Generating realistic graphs with deep auto-regressive models. In International Conference on Machine Learning. PMLR, 57085717.Google ScholarGoogle Scholar
  255. [255] Neil Daniel, Segler Marwin, et al. 2018. Exploring deep recurrent models with reinforcement learning for molecule design. In (ICLR’18). Workshop track.Google ScholarGoogle Scholar
  256. [256] Ahn Sungsoo, Kim Junsu, et al. 2020. Guiding deep molecular optimization with genetic exploration. Conference on Neural Information Processing Systems 33 (2020), 1200812021.Google ScholarGoogle Scholar
  257. [257] Zhou Zhenpeng, Kearnes Steven, et al. 2019. Optimization of molecules via deep reinforcement learning. Scientific Reports 9, 1 (2019), 110.Google ScholarGoogle Scholar
  258. [258] Winter Robin, Montanari Floriane, et al. 2019. Efficient multi-objective molecular optimization in a continuous latent space. Chemical Science 10, 34 (2019), 80168024.Google ScholarGoogle ScholarCross RefCross Ref
  259. [259] Aumentado-Armstrong Tristan. 2018. Latent molecular optimization for targeted therapeutic design. arXiv preprint arXiv:1809.02032 (2018).Google ScholarGoogle Scholar
  260. [260] Fu Tianfan, Xiao Cao, et al. 2021. Mimosa: Multi-constraint molecule sampling for molecule optimization. In AAAI Conference on Artificial Intelligence, Vol. 35. 125133.Google ScholarGoogle ScholarCross RefCross Ref
  261. [261] Orengo Christine A., Todd Annabel E., and Thornton Janet M.. 1999. From protein structure to function. Current Opinion in Structural Biology 9, 3 (1999), 374382.Google ScholarGoogle ScholarCross RefCross Ref
  262. [262] Jumper John, Evans Richard, et al. 2021. Highly accurate protein structure prediction with AlphaFold. Nature 596, 7873 (2021), 583589.Google ScholarGoogle ScholarCross RefCross Ref
  263. [263] Wang Jingxue, Cao Huali, et al. 2018. Computational protein design with deep learning neural networks. Scientific Reports 8, 1 (2018), 19.Google ScholarGoogle Scholar
  264. [264] Ding Wenze, Nakai Kenta, and Gong Haipeng. 2022. Protein design via deep learning. Briefings in Bioinformatics 23, 3 (2022), bbac102.Google ScholarGoogle ScholarCross RefCross Ref
  265. [265] Zhan Fangneng, Yu Yingchen, et al. 2021. Multimodal image synthesis and editing: A survey. arXiv preprint arXiv:2112.13592 (2021).Google ScholarGoogle Scholar
  266. [266] Taylor Paul. 2009. Text-to-Speech Synthesis. Cambridge University Press.Google ScholarGoogle ScholarCross RefCross Ref
  267. [267] Ning Yishuang, He Sheng, et al. 2019. A review of deep learning based speech synthesis. Applied Sciences 9, 19 (2019), 4050.Google ScholarGoogle ScholarCross RefCross Ref
  268. [268] Fooshee David, Mood Aaron, et al. 2018. Deep learning for chemical reaction prediction. Molecular Systems Design & Engineering 3, 3 (2018), 442452.Google ScholarGoogle ScholarCross RefCross Ref
  269. [269] Schwaller Philippe, Vaucher Alain C., et al. 2021. Prediction of chemical reaction yields using deep learning. Machine Learning: Science and Technology 2, 1 (2021), 015016.Google ScholarGoogle ScholarCross RefCross Ref
  270. [270] Cova Tânia F. G. G. and Pais Alberto A. C. C.. 2019. Deep learning for deep chemistry: Optimizing the prediction of chemical patterns. Frontiers in Chemistry 7 (2019), 809.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Controllable Data Generation by Deep Learning: A Review

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Computing Surveys
        ACM Computing Surveys  Volume 56, Issue 9
        September 2024
        980 pages
        ISSN:0360-0300
        EISSN:1557-7341
        DOI:10.1145/3613649
        • Editors:
        • David Atienza,
        • Michela Milano
        Issue’s Table of Contents

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 25 April 2024
        • Online AM: 9 March 2024
        • Accepted: 31 January 2024
        • Revised: 6 September 2023
        • Received: 12 November 2022
        Published in csur Volume 56, Issue 9

        Check for updates

        Qualifiers

        • survey
      • Article Metrics

        • Downloads (Last 12 months)255
        • Downloads (Last 6 weeks)163

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Full Text

      View this article in Full Text.

      View Full Text