How is AI accelerating Materials Discovery—and why are GFlowNets a core ingredient ?

The challenge of materials discovery
Discovering new materials is a critical prerequisite for scaling and inventing next-generation technologies. From clean energy to electrode materials for batteries, industrial sectors rely on materials that meet increasing performance, cost, and sustainability targets. Yet, despite their pivotal role, the discovery of new materials continues to be one of the most intricate and multidimensional problems in science.
One reason is that the number of possible materials is exponentially large, with some estimates suggesting that there could be over 10⁶⁰ stable compounds. Even with advanced high-throughput labs and simulation software, navigating this space manually is like looking for a needle in a massive haystack. On top of that, chemists usually do not just seek materials with a single desirable property but rather a combination of many aspects, such as cost, performance, and environmental impact. This level of complexity far exceeds what humans can feasibly manage alone, making it a domain where Artificial Intelligence (AI) can play a transformative role.
How can AI help discover new materials?
Over the past two years, AI has made significant strides in materials science, with major actors like Google DeepMind, Microsoft, and Meta releasing high-impact research that changes how we explore the chemical space.
In 2023, researchers at Google DeepMind managed to generate millions of theoretically stable materials with their model called GNoME [1] (published in Nature). They achieved it by training a neural network to predict the stability of new inorganic crystals created by (randomly) mutating known ones. This predictive model was then refined in a process called active learning, where the model selected the most informative candidates and evaluated them using quantum simulations to further improve the accuracy of their predictions . But while GNoME used AI purely to estimate the stability of these new candidate materials, models like MatterGen [2] from Microsoft Research (published in Nature earlier this year) go one step further: rather than exploring randomly around inorganic crystals known to chemists, they leverage AI to generate, from scratch, new materials that match target properties such as a specific band gap.
These advances point to a powerful idea: AI can perform an intelligent first identification of candidates in the chemical space, concentrating on promising regions instead of searching completely randomly. This is essential because each evaluation has a cost. Whether it’s a quantum simulation like density functional theory (DFT) or a real-world lab experiment, every evaluation demands significant time and resources. Therefore, success should not be measured only in terms of the quantity of materials being screened, but in the way they are prioritized for deeper investigation.
These ML models are usually only as good as the data on which they are trained. That’s why foundational datasets like LeMat-Bulk & LeMat-Traj [3], which we recently released in collaboration with HuggingFace, are so important. By unifying and deduplicating millions of quantum chemistry results from databases like Materials Project, OQMD, and Alexandria, LeMaterial series provides a clean and consistent foundation for training the next generation of generative AIs.
However, chemists and material scientists often still perceive AI as a black box, undoubtedly powerful but opaque. This does not foster trust and usage, especially as purely data-driven models can sometimes generate candidates that violate basic chemical intuitions, which experts would spot immediately. That’s why at Entalpic, we focus on building a generative AI that is transparent, controllable, and grounded in chemistry. Our objective is to empower domain experts, not replace them.
Building materials, step-by-step
A notable advance in generative AI for material discovery is a family of models called Generative Flow Networks, or GFlowNets for short. Initially developed by the 2018 A. M. Turing Award laureate Yoshua Bengio and his team at Mila, Québec’s AI Research Institute in Montréal, GFlowNets take a fundamentally different approach to crystal generation. Rather than generating a material all at once, just like MatterGen generating a crystal or even ChatGPT writing some text, a GFlowNet builds it step-by-step, following a structured and principled process that mirrors the way scientists might approach the same task.
Building upon a GFlowNet model called Crystal-GFN [4] designed specifically to create crystalline materials, generating a new crystal would typically involve four different steps: (1) selecting a crystallographic space group (i.e., the symmetries of the crystal), (2) choosing the atomic composition and the relative atom positions in space, (3) defining the lattice parameters (i.e., the lengths and angles of the crystal), and (4) evaluating how stable or performant the structure is for a certain application. This step-by-step approach allows experts to inspect and control every stage of the generation process. In particular, it makes it possible to incorporate key constraints, such as symmetry compatibility, charge neutrality, or realistic atomic densities, and ensure that the materials generated are valid and stable.
Figure 1. Simplified schematic of crystal generation process in Crystal-GFN
But beyond mere stability, which is only one piece of the puzzle when it comes to discovering new materials, GFlowNets let us guide generation towards practical goals. They can bias the search towards synthesizable, rare-earth-free, or CO2-efficient materials, to name a few, and even combinations of these characteristics. Furthermore, GFlowNets generate diverse sets of high-potential candidates, and not just the best guess. This is crucial in industries where trade-offs are constant and the right material often lies off the beaten path.
It’s important to recognize that AI-generated materials constitute only the initial phase of a longer process. Like any hypothesis, they need to be tested via quantum simulations and eventually in the lab. But GFlowNets help us generate candidates rooted in chemistry, tailored to industrial constraints, and ready for validation.
AI as a partner for chemists
AI is not meant to replace intuition—it’s here to expand it. Tools like GFlowNets are designed to explore vast regions of the chemical space that would otherwise remain out of reach for practitioners. They support chemists in navigating complex trade-offs involving cost, performance, scalability, and environmental impact, all while preserving full control over every step of the generation process to adapt it to specific industrial requirements.
If you are in industrial R&D, this is the time to engage. Not just to adopt AI, but to shape it. To help build models that reflect real-world challenges and accelerate the development of sustainable, high-performance materials.
Learn more
If you are eager to go deeper, explore these recent resources:
- Merchant, A., Batzner, S., Schoenholz, S.S. et al. ”Scaling deep learning for materials discovery.” Nature 624, 80–85 (2023). https://doi.org/10.1038/s41586-023-06735-9
- Zeni, C., Pinsler, R., Zügner, D. et al. ”A generative model for inorganic materials design.” Nature 639, 624–632 (2025). https://doi.org/10.1038/s41586-025-08628-5
- LeMaterial https://lematerial.org
- Mila AI4Science, et al. “Crystal-GFN: sampling crystals with desirable properties and constraints.” arXiv preprint arXiv:2310.04925 (2023).
Stay tuned; we’ll continue sharing more about how AI is reshaping material discovery and how your industry can lead this change.

Written by Tristan Deleu, Senior ML Engineer & key researcher in GFlowNets.
Contact the Entalpic team at contact@aqua-tapir-413535.hostingersite.com

Written by Tristan Deleu, Senior ML Engineer & key researcher in GFlowNets.
Contact the Entalpic team at contact@aqua-tapir-413535.hostingersite.com