Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Generative Artificial Intelligence (GAI) is attracting the increasing attention of materials community for its excellent capability of generating required contents. With the introduction of Prompt paradigm and reinforcement learning from human feedback (RLHF), GAI shifts from the task-specific to general pattern gradually, enabling to tackle multiple complicated tasks involved in resolving the structure-activity relationships. Here, we review the development status of GAI comprehensively and analyze pros and cons of various generative models in the view of methodology. The applications of task-specific generative models involving materials inverse design and data augmentation are also dissected. Taking ChatGPT as an example, we explore the potential applications of general GAI in generating multiple materials content, solving differential equation as well as querying materials FAQs. Furthermore, we summarize six challenges encountered for the use of GAI in materials science and provide the corresponding solutions. This work paves the way for providing effective and explainable materials data generation and analysis approaches to accelerate the materials research and development.
Liu Yue, Zhao Tianlu, Ju Wangwei, Shi Siqi. Materials discovery and design using machine learning. J. Materiom. 2017;3:159-77.
Liu Yue, Guo Biru, Zou Xinxin, Li Yajie, Shi Siqi. Machine learning assisted materials design and discovery for rechargeable batteries. Energy Storage Mater 2020;31:434-50.
Jovanović Mlađan, Campbell Mark. Generative artificial intelligence: trends and prospects. Computer 2022;55:107-12.
Cao Hanqun, Tan Cheng, Gao Zhangyang, Chen Guangyong, Ann Heng Pheng, Stan Z Li. A survey on generative diffusion model. arXiv:220902646 2022.
Jabbar Abdul, Li Xi, Omar Bourahla. A survey on generative adversarial networks: variants, applications, and training. ACM Comput Surv 2021;54:1-49.
Sanchez-Lengeling Benjamin, Aspuru-Guzik Alán. Inverse molecular design using machine learning: generative models for matter engineering. Science 2018;361:360-5.
Luo Shitong, Shi Chence, Xu Minkai, Tang Jian. Predicting molecular conformation via dynamic graph score matching. Adv Neural Inf Process Syst 2021;34:19784-95.
Chen Litao, Zhang Wentao, Nie Zhiwei, Li Shunning, Pan Feng. Generative models for inverse design of inorganic solid materials. J. Mater. Inform. 2021;1:4.
Zhao Yong, Edirisuriya M, Siriwardane Dilanga, Wu Zhenyao, Fu Nihang, Al-Fahdi Mohammed, Hu Ming, et al. Physics guided deep learning for generative design of crystal materials with symmetry constraints. npj Comput Mater 2023;9:38.
Liu Pengfei, Yuan Weizhe, Fu Jinlan, Jiang Zhengbao, Hayashi Hiroaki, Graham Neubig. Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput Surv 2023;55:1-35.
Wang Yaqing, Yao Quanming, James T, Kwok. Generalizing from a few examples: a survey on few-shot learning. ACM Comput Surv 2020;53(3):1-34.
Xian Yongqin, Christoph H. Lampert, Bernt Schiele, and Zeynep Akata. Zero-shot learning—a comprehensive evaluation of the good, the bad and the ugly. IEEE Trans Pattern Anal Mach Intell 2018;41:2251-65.
Ouyang Long, Wu Jeffrey, Jiang Xu, Almeida Diogo, Wainwright Carroll, Mishkin Pamela, et al. Training language models to follow instructions with human feedback. Adv Neural Inf Process Syst 2022;35:27730-44.
Goodfellow Ian, Pouget-Abadie Jean, Mirza Mehdi, Xu Bing, Warde-Farley David, Ozair Sherjil, et al. Generative adversarial networks. Commun ACM 2020;63:139-44.
Gulrajani Ishaan, Ahmed Faruk, Martin Arjovsky, Vincent Dumoulin, Aaron C Courville. Improved training of wasserstein GANs. Adv Neural Inf Process Syst 2017;30.
Xu Lei, Skoularidou Maria, Cuesta-Infante Alfredo, Veeramachaneni Kalyan. Modeling tabular data using conditional GAN. Adv Neural Inf Process Syst 2019;32.
Ma Boyuan, Wei Xiaoyan, Liu Chuni, Ban Xiaojuan, Huang Haiyou, Wang Hao, et al. Data augmentation in microscopic images for material data mining. npj Comput Mater 2020;6:1-9.
Marani Afshin, Jamali Armin, Moncef L, Nehdi. Predicting ultra-high-performance concrete compressive strength using tabular generative adversarial networks. Materials 2020;13:4757.
Yang Zhiyuan, Li Shu, Li Shuai, Jia Yang, Liu Dongrong. A two-step data augmentation method based on generative adversarial network for hardness prediction of high entropy alloy. Comput Mater Sci 2023;220:112064.
Dan Yabo, Zhao Yong, Xiang Li, Li Shaobo, Hu Ming, Hu Jianjun. Generative adversarial networks (GAN) based efficient sampling of chemical composition space for inverse design of inorganic materials. npj Comput Mater 2020;6:84.
Song Yuqi, Edirisuriya M, Siriwardane Dilanga, Zhao Yong, Hu Jianjun. Computational discovery of new 2D materials using deep learning generative models. ACS Appl Mater Interfaces 2021;13:53303-13.
Kim Sungwon, Noh Juhwan, Gu Geun Ho, Aspuru-Guzik Alan, Jung Yousung. Generative adversarial networks for crystal structure prediction. ACS Cent Sci 2020;6:1412-20.
Long T, Fortunato NM, Opahle I, et al. Constrained crystals deep convolutional generative adversarial network for the inverse design of crystal structures. npj Comput Mater 2021;7:66.
Michelucci U. An introduction to autoencoders. arXiv preprint arXiv:220103898 2022.
Chen X, Kingma DP, Salimans T, et al. Variational lossy autoencoder. arXiv preprint arXiv:161102731 2016.
Vahdat A, Kautz J. NVAE: a deep hierarchical variational autoencoder. Adv Neural Inf Process Syst 2020;33:19667-79.
Van Den Oord A, Vinyals O. Neural discrete representation learning. Adv Neural Inf Process Syst 2017;30.
Dupont E. Learning disentangled joint continuous and discrete representations. Adv Neural Inf Process Syst 2018;31.
Razavi A, Van den Oord A, Vinyals O. Generating diverse high-fidelity images with vq-vae-2. Adv Neural Inf Process Syst 2019;32.
Chen RT, Li X, Grosse RB, et al. Isolating sources of disentanglement in variational autoencoders. Adv Neural Inf Process Syst 2018;31.
Gomez-Bombarelli R, Wei JN, Duvenaud D, et al. Automatic chemical design using a data-driven continuous representation of molecules. ACS Cent Sci 2018;4(2):268-76.
Noh J, Kim J, Stein HS, et al. Inverse design of solid-state materials via a continuous representation. Matter 2019;1(5):1370-84.
Oubari F, De Mathelin A, Décatoire R, et al. A binded VAE for inorganic material generation. arXiv preprint arXiv:211209570 2021.
Ho J, Jain A, Abbeel P. Denoising diffusion probabilistic models. Adv Neural Inf Process Syst 2020;33:6840-51.
Dhariwal P, Nichol A. Diffusion models beat GANs on image synthesis. Adv. Neural Inform. Process. Sys. 34 2021;34(Neurips 2021):8780-94.
Song J, Meng C, Ermon S. Denoising diffusion implicit models. arXiv preprint arXiv:201002502 2020.
Watson D, Ho J, Norouzi M, et al. Learning to efficiently sample from diffusion probabilistic models. arXiv preprint arXiv:210603802 2021.
Vahdat A, Kreis K, Kautz J. Score-based generative modeling in latent space. Adv Neural Inf Process Syst 2021;34:11287-302.
Austin J, Johnson DD, Ho J, et al. Structured denoising diffusion models in discrete state-spaces. Adv Neural Inf Process Syst 2021;34:17981-93.
Kingma D, Salimans T, Poole B, et al. Variational diffusion models. Adv Neural Inf Process Syst 2021;34:21696-707.
Song Y, Durkan C, Murray I, et al. Maximum likelihood training of score-based diffusion models. Adv Neural Inf Process Syst 2021;34:1415-28.
Bao F, Li C, Zhu J, et al. Analytic-dpm: an analytic estimate of the optimal reverse variance in diffusion probabilistic models. arXiv preprint arXiv:220106503 2022.
Kingma Durk P, Dhariwal Prafulla. Glow: generative flow with invertible 1x1 convolutions. Adv Neural Inf Process Syst 2018;31:10236-45.
Chen Ricky TQ, Rubanova Yulia, Bettencourt Jesse, Duvenaud David K. Neural ordinary differential equations. Adv Neural Inf Process Syst 2018;31:6572-83.
Ohno Hiroshi. Training data augmentation: an empirical study using generative adversarial net-based approach with normalizing flow models for materials informatics. Appl Soft Comput 2020;86:105932.
Radford Alec, Wu Jeffrey, Child Rewon, Luan David, Amodei Dario, Sutskever Ilya. Language models are unsupervised multitask learners. OpenAI blog 2019;1:9.
Brown Tom, Mann Benjamin, Ryder Nick, Subbiah Melanie, Kaplan Jared D, Dhariwal Prafulla, et al. Language models are few-shot learners. Adv Neural Inf Process Syst 2020;33:1877-901.
Hong Zijian. ChatGPT for Computational Materials Science: A Perspective. Energy Mater Adv. 2023;4:0026.
Chen Xing, Abreu Araujo Flavio, Riou Mathieu, Jacob Torrejon, Ravelosona Dafiné, Wang Kang, et al. Forecasting the outcome of spintronic experiments with neural ordinary differential equations. Nat Commun 2022;13:1016.
Li Yajie, Sha Liting, Lv Peili, Qiu Na, Zhao Wei, Chen Bin, et al. Influences of separator thickness and surface coating on lithium dendrite growth: a phase-field study. Materials 2022;15:7912.
Liu Yue, Zou Xinxin, Yang Zhengwei, Shi Siqi. Machine learning embedded with materials domain knowledge. J Chin Ceram Soc 2022;50:863-76.
Fuhr Addis S, Sumpter Bobby G. Deep generative models for materials discovery and machine learning-accelerated innovation. Frontiers in Materials 2022;9:865270.
Liu Yue, Yang Zhengwei, Zou Xinxin, Ma Shuchang, Liu Dahui, Avdeev Maxim, et al. Data quantity governance for machine learning in materials science. Natl Sci Rev 2023;10. nwad125.
Shi Siqi, Sun Shiyu, Ma Shuchang, Zou Xinxin, Qian Quan, Liu Yue. Detection method on data accuracy incorporating materials domain knowledge. J Inorg Mater 2022;37:1311.
Weston L, Tshitoyan V, Dagdelen J, Kononova O, Trewartha A, Persson KA, et al. Named entity recognition and normalization applied to large-scale information extraction from the materials science literature. J. Chem. Inf. Model. 2019;59:3692-702.
Tshitoyan Vahe, Dagdelen John, Weston Leigh, Dunn Alexander, Rong Ziqin, Kononova Olga, et al. Unsupervised word embeddings capture latent knowledge from materials science literature. Nature 2019;571:95-8.
Nie Zhiwei, Zheng Shisheng, Liu Yuanji, Chen Zhefeng, Li Shunning, Lei Kai, et al. Automating materials exploration with a semantic knowledge graph for Li-ion battery cathodes. Adv Funct Mater 2022;32:2201437.
Liu Yue, Ge Xianyuan, Yang Zhengwei, Sun Shiyu, Liu Dahui, Avdeev Maxim, et al. An automatic descriptors recognizer customized for materials science literature. J Power Sources 2022;545:231946.
Liu Yue, Ding Lin, Yang ZhengWei, Ge XianYuan, Liu DaHui, Liu Wei, et al. Domain knowledge discovery from abstracts of scientific literature on Nickel-based single crystal superalloys. Sci China Technol Sci 2023.
Ishikawa Atsushi. Heterogeneous catalyst design by generative adversarial network and first-principles based microkinetics. Sci Rep 2022;12:11657.
Shen Yujun, Yang Ceyuan, Tang Xiaoou, Zhou Bolei. Interfacegan: interpreting the disentangled face representation learned by GANs. IEEE Trans Pattern Anal Mach Intell 2020;44:2004-18.
Liu Yue, Liu Zitu, Li Shuang, Yu Zhenyao, Guo Yike, Liu Qun, et al. Cloud-VAE: variational autoencoder with concepts embedded. Pattern Recogn 2023;140:109530.
Zhao Yong, Al-Fahdi Mohammed, Hu Ming, Edirisuriya M, Siriwardane D, Song Yuqi, Nasiri Alireza, et al. High-throughput discovery of novel cubic crystal materials using deep generative neural networks. Adv Sci 2021;8:e2100566.
Ren Pengzhen, Xiao Yun, Chang Xiaojun, Huang Po-yao, Li Zhihui, Chen Xiaojiang, et al. A comprehensive survey of neural architecture search: challenges and solutions. ACM Comput Surv 2021;54:1-34.
Chen Chun-Teh, Gu Grace X. Generative deep neural networks for inverse materials design using backpropagation and active learning. Adv Sci 2020;7:1902607.
This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).