🔬Why There Is No "AlphaFold for Materials" — AI for Materials Discovery with Heather Kulik
Heather Kulik, MIT professor of chemical engineering, discusses the intersection of AI and materials discovery, explaining why there's no 'AlphaFold for materials' yet. She covers her work using machine learning to accelerate materials discovery, including a breakthrough in creating polymers four times tougher through AI-discovered quantum mechanical phenomena.
- AI in materials science is most powerful for solving multidimensional optimization problems with 7+ objectives, providing 100-1000x speedup per dimension
- Current machine learning models for materials often fail catastrophically when applied to real problems, with foundation models being only 5x faster than traditional methods
- The biggest bottleneck in materials discovery is the interface between computational predictions and experimental validation
- LLMs have Wikipedia-level chemistry knowledge but fail at basic tasks like designing molecules with specific atom counts
- Materials science lacks the experimental ground truth datasets that enabled AlphaFold's success in protein folding
"The thing I constantly do every time an LLM is updated is I just ask it, please design me a ligand that has 22 atoms. I can never get an answer that has 22 atoms."
"ChatGPT is super good at Wikipedia level chemistry knowledge."
"If I could just give up ever doing a DFT calculation again and just rely on machine learned potentials and if they were two orders of magnitude faster than the traditional approach, that would change how we're doing science."
"You should learn chemistry well enough to know when these models are right or wrong. And if you don't know any chemistry at all, it's hard to know if you're assessing correctly."
There's a school of thought that why should I bother to learn chemistry or physics or whatever when ChatGPT has PhD level understanding of that?
0:00
Anyway, ChatGPT is super good at Wikipedia level chemistry knowledge. I'm really interested in molecular design. Like, how do you find a new ligand that can go into a transition metal complex? And what that means is that some combination of atoms and it's going to bind to the metal and it's going to change its properties. The thing I constantly do every time an LLM is updated is I just ask it, please design me a ligand that has 22 atoms. I can never get an answer that has 22 atoms.
0:12
All right, we're really excited to have Heather Kulik here. She's a professor of chemical engineering at mit. Heather has done some amazing work in material science and computational chemistry. Um, but we're particularly excited to have her today because she has for almost her entire career been working on the intersection of using data driven methods, AI and using applying them to improve materials and understanding materials. And she has a lot of, like, really interesting opinions about what works and how do you approach these problems to get the most out of them. So, yeah, we're really excited to have you here and yeah, maybe to get started, can you just tell us about like one of the coolest things you've done in Europe and in for a kind of AI engineering audience?
0:45
Yeah. So my, my group, we work a lot in accelerated discovery of new materials. When I first started out, we were just really using AI to make predictions we'd normally make with computational models, just make them faster. But the question I would often get when we were doing that was, okay, but what's surprising, what's sort of something from AI that I wouldn't have already known if I were a really smart chemist or a really smart material scientist. And you make all these computational predictions. Has anyone actually made in the lab something that you predicted? Recently I was able to do a really nice demonstration where the answer to both of those questions was very clear from the work. So we were able to scale screen with artificial intelligence a set of thousands, tens of thousands of materials, where each individual experiment, if it were done in the lab, would have taken months to years. And through AI, we uncovered this sort of unexpected chemical phenomenon that led to a emergent property in what's known as a polymer network. So plastics that would make the polymer about four times tougher. And when we showed the design that AI had come up with to the experimentalists, they were really surprised. They would have never come on this, on their own. And then we were able to convince them to make it in the lab. And in fact, it was this tougher material. And where this has applications is if we can make plastics tougher, then we can get more use out of them and it'll ultimately address some of the problems we have with overall durability and use of plastics. So I think that's an example of some of the promise of AI and materials discovery.
1:36
Cool. So can you dig into that a little bit? What was the surprising chemical discovery there?
3:23
So it's sort of hard for me to think about how to explain it without getting too deep into the chemistry. But basically these are molecules that have to break apart. And when they break apart, they make the overall structure that they're in tougher. So a little part of the material breaks and that helps to dissipate the force. Normally the way you would think about making it easier to break apart these small molecular components might be to create a hinge so they can kind of peel open instead of sliding apart. But what we discovered was that there was a fully quantum mechanical phenomenon. There was really no way for us to predict this, you know, based on anything else where the electrons just move around in a different way. So that at this moment where the molecule is going to break apart, it's a lot more stabilized. These types of concepts, they're sort of similar to what's kind of known about how catalysts and enzymes work. But it had never before been shown in these polymer materials.
3:30
So this is sort of like the fuse in the Bay Bridge that sort of like allows the bridge to keep its structural integrity during an earthquake by having a controlled brake. Is that kind of.
4:30
Yeah, yeah. So we weren't the first ones to discover that phenomenon on its own. The general phenomenon that putting little places that could break to make the network stronger, that was published in Science magazine a couple years ago. But the specific way we came up with to design the material to do this, that was. That was our new contribution.
4:43
How did you. You mentioned that, you know, you started off in accelerating kind of existing methods using sort of enhanced computation. What caused you to take that leap to more machine learning based methods?
5:02
So, you know, I was drawn to data driven discovery pretty early on, sort of before I even knew the phrase machine learning. And I guess I was just really excited by what you could learn from patterns and data. Back then we were trying to call it chem informatics and just sort of trying to think about, you know, in what ways could you unearth trends in data? Because I. I started my career actually working kind of one molecule at a time or one material at a time. And I was just impatient. I wanted to be able to sort of understand not just one molecule at a time and write one paper about it, which is something people would have been happy to do back when I was starting my career in the mid-2000s, but to actually kind of unearth broader trends in how you understand how material is going to behave. Somewhere around 2015, 2016, I realized it was a bad idea to call things cheminformatics, and it was a good idea to start calling things machine learning. And I had a brilliant student, John Paul Genet, who is now, I think, an assistant director at AstraZeneca in Sweden, running their inverse design program. He and I originally talked about all sorts of ways of thinking about materials design, and he very quickly adapted that into training neural networks. And that's sort of when I thought we were in the first sort of hype cycle, the first wave. But I think compared to what's going on right now, it was a tiny baby wave.
5:17
I read in your paper that that was actually a class project or something.
6:55
Yeah, yeah, that's right. You know, he just said, I have to do something for my homework. And that's how we. That's how we got into it. Got into it.
6:59
I've also read in your paper that you've done a lot of work, like slightly more recently on active learning and using. Can you talk a little bit about that?
7:06
Yeah, yeah. So even that polymer example I was giving, that would have been active learning in principle, but we sort of stopped after one generation because. Because we had exhausted the space. But, but I think one of the areas where machine learning kind of just with what's out there right now, has the most promising chemical sciences is in solving multidimensional challenges. So right now we're working on a project in metal organic frameworks where we're trying to solve trade offs relevant for direct capture of CO2 from the air. And so in order to find a material that's good for that, we would worry about its cost, its stability in, say, aqueous, humid environments, its ability to take in CO2 over other molecules, its mechanical stability. Is it going to hold up under force? Is it thermally stable? Can you heat it up and will it be okay? I'm just naming a few, but in total, right now, in an active learning campaign, we're working on seven different objectives. And, and usually just even for a not so accurate machine learning model, you get at least 100 to 1000 fold speed up for every dimension you're optimizing over. So the real promise is going to be in searching for that needle in a haystack with say, seven objectives and doing something where you're not waiting for the models to be accurate before you start doing that optimization. That's really the promise of active learning. Yeah.
7:14
That has an interesting parallel in my mind to the pharma world, where you have a lot of computation work in the discovery process, but that actually getting the drug out to in people's hands is often the bottleneck for a drug.
8:48
And also, you know, what happens to the drug when it sits on the shelf for three months, that kind of thing. Yeah.
9:04
Are these medical organic frameworks, what are the kinds of things that they're eastward for?
9:12
They're used most in gas storage, sensing and separations. They're used in combination with polymer composites. They have really strong promise for CO2 capture especially. But people have looked at them for catalysis. The limitation on catalysis has been how stable are they. So one of the things we've spent a lot of time on is trying to be able to predict their stability. But they're used for all sorts of things, even drug delivery. What they have the opportunity to do is really place precise chemical groups in specific orientations that can ultimately allow for what's known as host guest interaction. So basically kind of create a glove to have a targeted interaction with a guest molecule in the metal organic framework.
9:18
I see. And just for the non chemists, metal organic framework Legos for chemistry, is that.
10:10
Yeah, yeah. Metal organic frameworks, I think are going to be a little bit more of a household name among some engineers because the discoverers of those materials just won the Nobel Prize in chemistry this year. So as much as that can make something in chemistry a household name. But they're basically like Tinker toys or Legos and they have different building blocks that can be combined in basically infinite ways to create very precise chemistry.
10:19
I see.
10:47
Maybe for context, could we see if I can, like, what are the techniques you were using before you started? Or maybe in parallel with machine learning? And how does machine learning help you advance those? Like, what are the roles of the two?
10:48
So I started my career studying what's known as transition metal catalysis. If you look at the periodic table, the middle of it contains a bunch of metals. A good example would be iron. And all of those things sitting in the middle of the periodic table, they have what's referred to as an open shell. So the Electrons in those materials are not paired and they're as a result more reactive normally the way that you understand how they're going to behave. So for instance, they give rise to, you know, different combinations of these metals, give rise to the catalysts that are used in a large number of transformations, including the things that, say, feed and sustain most of the world's population, such as the Haber Bosch process for ammonia synthesis. And the way going back 20, 30, 50 years that people understood these materials and could enable their rational design is through quantum mechanical modeling. Quantum mechanical modeling, by using approximations to the Schroinger equation, it can be very accurate, but it's very computationally costly. And so a single quantum mechanical prediction, depending on the level of fidelity used, could take hours to days to weeks. And that's what I would have normally been doing before I got started in AI. Some of what we do these days is accelerating those quantum mechanical predictions as well as looking at, you know, an area that I'm particularly excited about is that not all quantum mechanical approximations are equal and you can actually use ML models to kind of predict what the best approximation to use is depending on the material studied.
11:02
Is that like closer is better or is it some. Is it. It's not really distance related in terms
12:48
of which method is the right method to use? Yeah. So we. It actually turns out to be quite complex. You can't just determine it from heuristics. So we actually in one area use the quantum account canonical wave function as inputs to neural networks to actually predict what is the right method to use and learn that mapping.
12:55
I see that's probably going to be.
13:15
That sounds like a challenge in the next but cool like 22 Adam Ligon challenge go.
13:16
I have a spicy question I want to ask. So there's a school of thought that why should I bother to learn chemistry or physics or whatever when ChatGPT as PhD level understanding of that anyway? And shouldn't I just focus on being really good at using AI for stuff? So I want to hear your thoughts.
13:24
My personal experience is that, and this will date itself immediately is that ChatGPT is super good at Wikipedia level chemistry knowledge. But one of my favorite things to actually throw at GPT as an anecdote is I'm really interested in molecular design. Like how do you find a new ligand that can go into a transition metal complex? And what that means is that some combination of atoms and it's going to bind to the metal and it's going to change its properties. And so the thing I Constantly do. Every time an LLM is updated is I just ask it, please design me a ligand that has 22 atoms. The first time I've done that, there are many ligands out there that have 22 atoms and I say I want it to bind to the metal with two nitrogen atoms. I can never get an answer that has 22 atoms. So then you can try a range and see how many times you can get a range. And so that's maybe a trivial thing, but that's something that a expert chemist could do in a second. So there are really good introductions to chemistry that I think you can get through conversations with an LLM, you can get a of insight into an area you're unfamiliar with. And for sure things have improved a lot. Like when I first tried typing in, you know, which exchange correlation functional should I use for this type of chemistry? The answers were completely wrong. They looked right, but they were completely wrong. I think things have gotten better because that knowledge is out there on the Internet and it's in the training data. But I think there's a lot of things that probably backing up a moment, you should learn chemistry well enough to know when these models are right or wrong. And if you don't know any chemistry at all, it's hard to know if you're assessing correctly. But I think that there are a lot of things that you don't have time to do a deep dive into that you can now get from say an LLM that can augment knowledge. But I think you have to start from somewhere and then use it as a tool rather than starting from zero and relying blindly on what an LLM will say. But one of my favorite things, if someone can get in one shot an LLM to generate me a 22 atom ligand, I would love to see it.
13:48
What do you think the biggest gaps that machine learning has from your experience that if you are an aspiring ML engineer with looking to take on a new problem from the machine learning side, what do you think someone could work on which would really help the chemistry side?
16:07
There are a lot of challenges out there where the data sets aren't large enough or diverse enough and so I think they've attracted less interest. The ones closest to my heart are reactivity predictions. So predicting which reactions will occur and why, especially in complex phenomena like in multiple elements and predicting those transformations. Another thing that I think there's not enough data on is just more diverse chemical bonding and more diverse chemistry. For me that's transition metals. But there's also questions of warm, dense Materials, sort of exotic phenomenon. We have really good data sets out there for really boring chemistry. So we have probably even if you're not a chemist, you're familiar with organic molecule data sets and organic molecules binding to proteins. Those are the common data sets out there. There's lots of challenges out there. The physics is much more complex and the things like how does matter behave when you shine light on it and you excite it into excited states? All sorts of things like that receive relatively little attention because there may not be a benchmark or a leaderboard yet for that. And so maybe it's on us chemists to generate more data sets. So those leaderboards are out there, but there's definitely a lot of interest in chemistry for which there has been less attention.
16:23
So in the protein world there's casp, right? And people have been working on that for a while and this led to AlphaFold, like kind of without CASP, AlphaFold probably wouldn't exist. Is there like an equivalent to CASP in the material science world?
17:57
So there are all sorts of repositories of fairly low fidelity DFT data on crystalline materials. So Materials Project, Open Catalyst Project, these do provide good leaderboards. But some of the limitations of that are the data comes from not very high fidelity, the density functional theory. So I'd say that's a second challenge is that all the smartest ML engineers right now are learning on data that is not going to be reflective of experiment. There aren't big experimental data sets. For example, one of the advantages of things like CASP is that it comes from an experimental ground truth, whereas that aspect just isn't available in materials.
18:11
As much as we talked about CASP and the role of CASP and AlphaFold, do you think that there is a problem, a way of phrasing this, that we could start collecting data at scale, that we could really have a community challenge which breaks open some open problem in your mind and maybe actually even stepping back beyond that, what would you want to have if there was like an alpha fold for materials, what would you want it to do?
18:56
One kind of murky area. So maybe I'm not going to directly answer this question. One murky area for us is electronic structure calculations are expensive and they should in principle give you the right answer. They should from first principles give you the right answer of how a material is going to behave. And a lot of people are scaling these up right now with machine learned inter atomic potentials on training data. And every time someone comes out with kind of a new data set trained On a. And they call it a foundation potential foundation model. It looks really good and then you get it into your lab and you say, okay, I want to use it for this problem I'm really excited about. And it starts doing kind of wacky things like molecules fall apart. I won't name names, but there was one that made a huge splash this summer and people started declaring, oh, this method is dead, this method is dead. We're all going to just use these neural network models now. It's only in my hands. The one I'm still not naming is only about five times faster than my fastest DFT calculation on a gpu. And it also doesn't work all the time. So I would say we need a more transparent way of trying to figure out if these models can really replace conventional physics based modeling. If they could. If I could just give up ever doing a DFT calculation again and just rely on machine learned potentials and if they were two orders of magnitude faster than the traditional approach, that would change how we're doing science. But there needs to be a little more rigor on what we consider just fitting data when that data maybe lacks quality. Or there needs to be a little bit tougher requirement for how we say this model can really replace the physics based modeling.
19:30
Yeah. So one of our theses is that the interface between bits and atoms is really the bottleneck where you have to the actually activity of trying things in the lab is the bottleneck. And you've addressed that to some extent in active learning, but I think that there's also an extent to which that just pure process and automation, good operational practice, those are important things. So that if you can push to automation on the one side, but on the other side that creates brittleness. So how do you think about kind of bridging that gap to experimental chemistry and using that sort of as a nature's computer to figure out things for your design process.
21:30
Yeah. So there are a lot of really smart people working in high throughput synthesis and experimentation and autonomous labs. I think the thing that that's interesting to me in that space at least, is that there are some types of experiments that at least as of the last conference I went to on this, are really hard for autonomous high throughput experimentation, but are really easy for a human and vice versa. And then there's, you know, the serendipity that a human might experience in the lab that a couple of people have tried to think about like, well, how do you, how do you introduce that noise into high throughput experimentation? So I think that's a challenge. Your question also brought to mind another point that I am by no means an expert on, but most people who actually work on getting materials to the device scale, say something that would be in your television or something like that, is they will tell you that it's not just the material, it's the process. And I think we're at, we're at ground zero. We're nowhere when it comes to like well how do we machine learn not just the structure and the properties, but also the role that processing plays. I don't think we know anything about how to do that.
22:23
Maybe for non experts, like with protein structure it's really easy to imagine, like oh, you can see these proteins and we can run some simulations and see them wiggling around and the structures look really pretty. What does the data look like for material science? There's the computations like DFT I think gives you something which looks like a crystal structure you can imagine. But then there's also like, is there experimental data where you can observe that crystal structure or is this mostly sort of like kind of probes where you're measuring individual properties which are kind of collective and with not fine grained.
23:41
So experimental structures are available. And the example I was giving is something we know is stable and we've seen a structure of it before and it will fall apart with some of these models. The challenge here is that what Alphafold has done really well is predict structures of globular proteins primarily with 20 natural amino acids. I could actually point to lots of cases where AlphaFold fails too for more interesting chemistry. The challenge is that you have a lot more than 20 building blocks when it comes to materials. And so there's lots of different ways to think about chemical bonding. And right now no potentials are really robustly encoding all of that bonding, especially with respect to metal, organic bonding.
24:14
Maybe a different way of saying it is with Alphafold. Alphafold is solving ground state structures. It's not looking at dynamics, which is I think consistent with some of your statements about needing quantum mechanics for catalytic enzymes. But you're saying even at just kind of ground state properties you're saying that just there are too many parameters and there's not like a clear set of interactions which is limited to a small number of building blocks.
25:00
The bonding is highly variable across all of material space. Now there's simple regions of material space. You can pick aluminum, aluminum is very boring. And you can write down, people in the 60s could write down on paper. You know how you need to model Aluminum, that's something that is pretty easy to fit a neural network potential to. But then if you want to get over to iron oxide, and then if you want to get over to high entropy alloys, there are definitely cases where people are using these methods. But I'd say a big challenge is that there's no real way to know if, when you go to bigger length scales and time scales, there's no real way to know if you're right or wrong. The experimental data is not there. Even interpreting, say looking at an image of an experiment surface, which you would want to do, it requires some degree of interpretation of that image. So it's just hard to know from experiment or from other computations if these types of models are correct. And they're certainly not correct across all of chemical space. And I'd say they could fail more catastrophically than alphafold obviously fails, though there are definitely failures of alphafold too.
25:27
Switching gears a little bit, I read in the new paper also that you had done some work with integrating textual information from papers into your. So it's kind of the AI that we all know and love right now. Can you talk about what kind of lift that gives the models and how did you actually do that? Integration?
26:42
Yeah, so we started, I guess about five years ago. So when we first started doing it, we were just doing sort of standard natural language processing and graph digitization. These days we use LLMs, but just to try to extract from the literature data sets of properties wherever people are widely reporting properties. And what we noticed is that there's a lot you can learn from these models. So you can, even on the scale of a few thousand data points, you can then do things like predict the temperature at which a moth will break apart based on experimental reports. But one of the funniest things I think we noticed is that you can get the temperature at which a material will break down two ways. One, you can get it from the graph, and two, you can get it from what the authors say about how they interpret the graph. And those two things do not line up. So people, you know, one of the challenges I think with literature extraction from papers is one would be the obvious mistakes people make. You know, no one's perfect, but the other would be just, you know, people interpret their results in different ways. And so if we're building models based on those interpretations, that's a challenge. In terms of LLMs, they've come a long way in terms of literature extraction, but they're still definitely sensitive to false positives. And I think the amount of Time we spend checking on LLMs to make sure that the data we're ingesting is accurate definitely is an overhead on those types of workflows.
27:01
I see. And what about the way that it might bias the discovery process? Right, because you have this known literature, your job as a chemist kind of sort of is to find new stuff. But if so, if you're emphasizing, if your computational method is pulling in literature, then maybe it's biasing you towards the previously reported results instead of something new.
28:31
You know, one of the ways we try to address that is we try to train a model on that literature, but then apply it to new structures that have never been seen before and try to really look at how far we can extend the model. But we are trying to answer this. In general, there are repositories out there of experimental data where you can have a sense of when it was published, what the structure is, what it was used for. And we're really trying to build generative models on top of that now to try to be able to say, well, if I know about the first 30 years of a field, can a model trained on that predict the next 20? I think that's an open question and what model is best at? You know, and maybe it won't get all of them, but maybe some of those discoveries that we think are new in the most recent 20 years, maybe, maybe some of them are trivial for a model to generalize to, whereas others are not. I think in an ideal case where we have the available literature data and we don't know, we could use uncertainty quantification to then identify, okay, these would be the most interesting materials to get into our data set.
28:54
I see in those data sets just for people who are interested in getting involved. So what are some of the, what's the best ones to get to get started with?
29:59
I don't know about the best. We've curated a few thousand data points of metal organic framework thermal stability, as well as metal organic framework activation stability, water stability. Other groups have curated other measures of stability. They're all out there, they're on our website, that kind of thing.
30:06
Awesome.
30:25
Do you imagine there being useful to create an initiative or a multi institutional funding a great source or something which really is trying to get data in a high throughput automated way? What would sort of like your dream be if you could organize something which, like, which really will drive the field forward in your mind?
30:26
I think the National Science foundation has one initiative. I've also heard about things with foundations before, sort of being interested in Getting putting together cloud labs so things that users can on demand make use of high throughput automation. I definitely think, I think having user facilities where a computational researcher like me could design an experiment and have it executed would be awesome. Having all that data collected in sort of a public way would be great. You know, the way that research right now gets published into papers, it's very hard to then extract back out. We spend a lot of energy trying to get it back out. And so some of this is a need also for maybe systematization of how results get reported so that they can be machine learning ready from day one when they're published. Some research sub fields are trying to do that, but it's not really developed across material science. But for sure I think there will be more shared facilities where people can make use of data from high throughput experimentation and that would be really awesome. I don't know if it'll come from companies donating equipment from National Science foundation or from, you know, private foundations.
30:49
Yeah, there is a large like philanthropic push in the biotech space. It seems like people haven't quite picked up on this as such an important field. Like, especially with things like materials for climate change. You can imagine in particular a very important problem that we could use a lot of cushions.
32:11
Yeah, that kind of brings up the question, there's been a ton of very recent materials investment for private companies, startups. Where does that leave in your mind the role of the academic in chemistry?
32:29
I ask myself that all the time or more recently in the past year. So in particular there's a lot of computer that companies have access to that academics don't. So I ask myself what can we do that's more creative, that doesn't require just brute force compute? And I think there is a lot of stuff that we can still do, but we have to ask those questions for sure. Microsoft Meta. Those ones are the companies that have basically infinite resources. And as an academic, I don't have infinite resources, you know, but we have an interest in problems that, you know, haven't crossed the radar of those companies yet. And I think as long as we, you know, whenever someone poses a problem to me now versus a few years ago, I try to make sure that we're not just in the process of trying to do something that you're throwing a lot of compute at, it would, would solve it.
32:45
Yeah, I think we're kind of running out of time, but would like to give you an opportunity. Call to action. What would you like our listeners to know about do? What should they do to get involved or something that you're really passionate about
33:50
I think I will stick to something kind of niche great so I think there is still a place for chemistry I will say that but my group develops a code for transition metal complex structure generation metal organic framework screening it's called mulsimplify or when we're working on MOFs we call it MOFsimplify. There's website versions of it that you can look up and not install anything but it's also on Conda and GitHub and if you do have an interest in transitionml complexes you know just try it out. It includes machine learning predictions but it also makes novel structures and I'm just really interested to hear ever if people are using it I know a lot of companies are using it but we sort of find out sort of after the fact so if you're interested more in this material space I'm definitely interested
34:05
and open to feedback Grateful awesome getting involved thank you very much Take care doctor.
34:54