Hello, and welcome back to the Cognitive Revolution!
Today I'm pleased to share a special crosspost from the a16z podcast, featuring a16z General Partner Anjney Midha – who also recently joined me on the Cognitive Revolution to discuss "Sovereign AI" – in conversation with Liam Fedus, former VP of post-training research and co-creator of ChatGPT at OpenAI, and Ekin Doğuş Çubuk, former head of materials science and chemistry research at Google DeepMind, who together have co-founded Periodic Labs and just announced a $300M investment led by a16z.
Before diving in, a quick note: while Turpentine was recently acquired by a16z, my editorial independence remains unchanged, and I'm sharing this episode simply because I think it offers a valuable perspective on the future of AI-powered science.
Regular listeners will no-doubt notice some overlap between this conversation and our recent episode with Radical AI – both companies believe that there simply isn't enough high-quality experimental data in the scientific literature to train foundation models for physics and chemistry, and both have raised serious capital to build automated physical laboratories that are meant to connect AI-generated hypotheses directly to physical experiments, using feedback from physical reality as the reinforcement learning signal, with the goal of teaching AI models a form of scientific intuition, and thereby accelerating scientific progress itself.
Of course, there are still many possible ways to focus such an ambitious company, and while Radical AI has recently announced a contract with the US Air Force to develop high-entropy alloys for use in hypersonic aviation, Periodic Labs has set the goal of discovering a high-temperature superconductor as their north star, with the expectation that to get there, they'll need to achieve countless sub-goals along the way, including autonomous synthesis and autonomous characterization.
Importantly, while the science and macro strategies are similar, the conversations are quite different.
While I tend to explore the technical details in arguably tedious depth, Anjney focused much more on the human and organizational dimensions of building an AI for Science company.
As you'll hear, because no human comes close to holding all the scientific knowledge and intuition they hope to train into their AI systems, Periodic Labs prioritizes people with intense curiosity and mission alignment, and does not require advanced degrees, takes pride in their "no stupid questions" culture, and hosts weekly teaching sessions in which ML researchers, physicists, and chemists can learn from one another.
And recognizing that $300M won't be enough to achieve their ultimate goals, and that even a wildly successful company is only one part of the broader scientific ecosystem, they have thoughtful plans to commercialize their progress in the form of an intelligence layer for advanced manufacturing companies, while also starting a grant program meant to elicit key contributions from academia.
Overall, I love the vision and ambition on display here, and admire the conviction with which a16z and others are backing it.
While this doesn't come up in the episode, I've long believed that long-term AI safety might best be achieved by creating domain-specific superintelligences, such that the AIs that advance fundamental science don't need to have advanced theory of mind or persuasion skills, and in any case, as much fun as I'm having playing around with Sora 2, it seems clear that a future of truly radical abundance requires AI systems that go beyond the digital environment and iterate against nature's own ground truth.
With that, I hope you enjoy this conversation about building an AI research company meant to develop systems that autonomously explore and deeply understand the physical world, from the a16z podcast, with host Anjney Midha, and Liam Fedus & Ekin Dogus Cubuk, founders of Periodic Labs.
Speaker 0: Ultimately, science is driven against experiment in the real world. And so that's what we're doing with Periodic Labs. We're taking these precursor technologies and we're saying, "Okay, if you care about advancing science, we need to have experiment in the loop."
Speaker 1: The applications of building an AI physicist, for lack of a better word, that can design the real world are so broad. You can apply them to advanced manufacturing. You can apply them to material science, to chemistry. Any process where there's R&D where the physical world's required, it seems like it will benefit from breakthroughs that Periodic is working on.
Speaker 2: For example, if you could find a 200 Kelvin superconductor, even before we make any product with it, to be able to see such quantum effects on, at such high temperatures, I think would be such an update to people's view of how they see the universe.
Speaker 1: So Liam, you were the co-creator of ChatGPT. Doge, you were running some of the physics teams at DeepMind. Let's talk about how you guys met, and what was the moment where you realized that you guys had to leave both of those labs to start Periodic?
Speaker 0: I believe we met eight years ago at Google Brain, flipping over a large tire.
Speaker 1: Yep.
Speaker 0: At the Google-
Speaker 1: You gotta give us more on that story.
Speaker 0: Um, so Google Rails, uh, was one of the gyms at Google Bra- uh, at the Google facilities. And I think that's where Doge and I met, and there was just this massive tire that, like, a single person basically can't flip single, like by themselves. And so Doge was trying to flip it, and he, like, pulled me over, and he's like, "Oh, I think the two of us could do it," so...
Speaker 1: And what, why were you trying to flip this tire? You know, just... Why is-
Speaker 0: No, why not?
Speaker 2: Um, but yeah, I- I tried doing it. I couldn't do it, and then I was like, "Who's the strongest person I can find?" And it was either Barrett or Liam. And I went with Liam. And it worked, we did flip it, so...
Speaker 0: Right.
Speaker 1: And was that, um, the moment where you guys both realized you had physics backgrounds? How did that happen? What, what, how, how, how did you go from, from flipping tires to flipping experiments?
Speaker 2: Yeah, I mean, so I don't know if Liam remembers this, but we would catch up, you know, over the years, and we would often end up talking about like quantum mechanics or like superconductivity. So it was like very common. But I never thought that we'd end up working on physics together. Um, so Liam was working on LLMs, and they were going really well. Um, and I was not using LLMs, but I was noticing that LLMs are becoming more and more impactful in my work. So one way it was becoming impactful is when I was trying to remember some things about chemistry, physics, I could just talk to the chatbot, and I actually learned a lot of stuff I forgot. Another way was, of course, coding. Like we were writing simulations, and the LLM was so helpful in writing these simulations for us. So then the question was, like, can we use LLMs kind of more as a first-class citizen in the physics research?
Speaker 0: Yeah, I think kinda leading up to this decision to leave, um, Doge and I were just, you know, connecting and talking about these different tech trees. W- we're looking at the improvements on language models, on reasoning. We're seeing what high compute reinforcement learning can do. And then on the material science side, we're seeing scaling laws within physics, within chemistry, uh, both with respect to simulations, with respect to experiment, and it's like the same kind of principles at play in ML. And I think to both of us, and to a lot of people in the field, the goal of this technology is accelerate science, accelerate physical R&D. It's, you know, chatbots was like a great milestone along the way, but we really wanna see technology out in the world.
Speaker 1: Mm-hmm.
Speaker 0: And we felt like this was just the right place to begin. Uh, physics is very verifiable. It's, it's a great reward function, fairly fast iteration loop. You have simulators for large classes of physical systems. And we felt like, in order to create this AI scientist, this is like the beginning of, of this path. So, built that conviction and then decided to found Periodic.
Speaker 1: Well, let's take a second to talk about what Periodic is, and what does it do?
Speaker 2: So, Periodic Labs is a frontier AI research lab that's trying to use LLMs to advance physics and chemistry. Um, we feel like having experiment in the loop tightly coupled with simulations in LLMs is extremely important. So, we're building up a lab that will generate high-throughput, high-quality data. And we will use LLMs and simulations in conjunction with the experiments to try to iterate. Science, by its nature, is a iterative, um, direction. And we feel like LLMs, using all these tools that are available to humans, can do a great job in accelerating physical R&D.
Speaker 0: I'd say the objective is, "Let's replace the reward function from math graders and code graders that we're using today." So like, math graders, you know, uh, to give an example, uh, you have a prompt, "What is two plus two?" You know the ground truth is four. You can put a lot of optimization pressure against problems like that, that are programmatically checkable. And what we're doing, and by having the lab, is we create a, a physically grounded reward function. That becomes the basis on which we're optimizing against. And so, if a simulator has some deficiencies or some issues, we always error correct, because for us, the ground truth is the experiment. Like, the RL environment... Nature, like, is our RL environment in, in our setting.
Speaker 1: Let's just take a second for folks who might not be familiar to explain what you guys mean by a lab that will verify RL in the real world. Can you talk a little bit about how experiments work? How do, how do, how are AI models trained today? And how is it, how are those different from how they're gonna be trained and, and developed and post-trained and deployed at Periodic? And it might be helpful to talk about how you created ChatGPT.
Speaker 0: So, ChatGPT originally, um, the technology evolved very rapidly over the, over the last few years.When we were first creating it, it was a very standard RLHF pipeline. So you have a pre-train model and it's sort of like just like raw substrate. And what you're trying to do is take this auto-completion model and turn it into something useful. The way we did it at that point was we would have supervised data. So given some input, we would say, "This is the desired output." So if we're trying to a- get it to act as an assistant, you know, we create some tuples like that. Then you run reinforcement learning, but now you're learning against a h- a reward function that's trained against human preferences. So humans will say, "Well, given this input, I would prefer completion A to completion B." And you do that over and over again, and you can create a reward function that can then be optimized against. That is sort of the basis of how we created ChatGPT. But then there's a huge gap between the original model and what we have today. And I think part of that is reasoning, but also part of that is just much better, more precise reward functions. So the reward functions that we were using originally couldn't determine whether you were mathematically correct or not. So early versions of ChatGPT were mathematically not particularly strong, and it sort of results from the reward function. What did you optimize against? You... The reward function basically encoded, be a friendly assistant, try to help people get to their thing. But it had no sense of, is this mathematically correct or not? Is this code valid or not? And we made huge advances over the correctness of our reward functions. Um, but this is all digital. Uh, we're creating tasks based on the internet, textbooks, papers, and this is great, this lays a foundation, but ultimately science is driven against experiment in the real world, and so that's what we're doing with periodic labs. We're taking these precursor technologies and we're saying, "Okay, if you care about advancing science, we need to have experiment in the loop, and that becomes our reward function for our agents." So as Dosh was saying, our agents are doing the same type of things you would use for coding or to help answer a query, but now instead of just giving tools like here's Python, here's a browser, now we have tools like quantum mechanics, so simulate different systems. But ultimately we're going to a lab, and then that becomes like the basis of what is the, the system optimizing against.
Speaker 1: So-
Speaker 0: So that's sort of just like the natural, like, end state of these systems.
Speaker 1: People in AI often say lab. Often what they're referring to is quite different from you, what you guys mean by a lab. H- Dosh, what, what's the difference?
Speaker 2: That's right. So as Dave mentioned, so far the LMs have gotten really good at logic and math. There's like verifiable rewards. What is, like, the next frontier in terms of, you know, inquiry after logic and math? I'd say it's physics. And then when you say physics, there are different energy scales. So there's astrophysics, studying galaxies, there's fusion, nuclear physics, but then there's the energy scale of physics that's more relevant to our life, and that's the quantum mechanics, like Schrodinger's equation. This is where, you know, biology happens, chemistry around us happens, materials happen. So we felt like our first lab should be basically probing that quantum mechanical energy scale. Um, and for us that would be physics at the level of solid state physics, material science, and chemistry. Uh, the kind of one of the more fundamental ways of making things around us is powder synthesis. So you take powders or existing materials, you mix them, and you heat them up to a certain temperature, and it becomes a new material. Uh, so that's one of our labs. We're gonna have a powder synthesis lab. And turns out this is one of those methods where robots can do it, like very cheap, simple methods. Um, I don't know if you saw this, uh, coffee-making robot in the SF airport. You know, a robot that's basically at that level can mix powders and put it in a furnace. Um, and that's a very rich field, so you can actually, using that method, discover new superconductors, magnets, you know, all kinds of materials that are very important for technologies around us. But at the core of it, it's just quantum mechanics. Uh, and we feel like teaching these LLMs to be foundation models but for quantum mechanics will be the next frontier for LLMs.
Speaker 1: Why haven't the models that are currently out in the world and deployed able to do this?
Speaker 2: Great question. I think, like, as Dave mentioned earlier, science is by its nature iterative, right? Like, even the smartest humans tried many times before they discovered the things they discovered. And I think maybe this is one of the confusing points about LLMs. LMS can be very smart, but if they're not iterating on science, they won't discover science. Or, you know, to be honest, humans won't either. Like, you put a human in a room without any chance to iterate on something, they won't discover anything important. So we feel like the important thing to teach these LLMs is the method of scientific inquiry. So you do simulations, you do theoretical calculations, you do experiments, you get results, and the results are probably incorrect or not what you wanted at first, but you iterate on it. And we feel like that hasn't been done yet. So this is what we want to do, but we feel like you have to do it with the real physics, not just the simulation. So this is why we have our own lab where the LLM will have the opportunity to iterate on its understanding of quantum mechanics.
Speaker 0: Fundamentally, machine learning models are good at what you train them to do.
Speaker 2: Right.
Speaker 0: And that's sort of, like, the nature of it. And so if a model is acting badly, you're like, "Well, did you train it to do that task?" Um, kind of building on Dosh's point, there's sort of like an epistemic uncertainty, this, like, reducible uncertainty that you aren't really building or collapsing unless you're actually running an experiment. So for instance, uh, one of the engineers on our team was looking at a reported property of some, like, some physical property in the literature, and it spanned many orders of magnitude. So if I train a system on that, these systems aren't magic, the best they can do is replicate that distribution. But it's really no closer to a deeper understanding of the universe, physics, chemistry. Um, then another point is it's very uncommon to publish negative results. All of the results are basically...... positive and a na- a valid negative result is very valuable. A negative result could be discarded 'cause you're, "Well, that is sloppy science." But there are, are valid negative results, and that's a learning signal. And this is something that our lab will produce as well. So I think these three things. There's just, like, noisy data, no negative results, and you need the ability to act in order to actually do science, which is an iterative endeavor. Those are, like, the core thesis of why we need a lab.
Speaker 1: And what might be the core way to measure if Periodic's progress against that goal in your guys' minds?
Speaker 2: One simple one is, let's say high temperature super conductivity. What is the highest temperature super conductor we synthesized? Uh, today, the best number for ambient pressure is 135 Kelvin or so, so we'll know very easily if we're doing well if we can go beyond that number. Um, so that's pretty fundamental. On the more applied side, you know, there's, uh, processing of materials and its effect on the material's properties. So we can just measure these properties directly. Let's say it's the ductility, it's the toughness, strength of the prop, uh, material. And as we measure it, the LLM will get very clear signal. It's hard to hack, you know, un- unless, unlike these other LLM training techniques. It's, like, really what you see in real life is the signal that's going to the LLM.
Speaker 0: Yeah, effectively. It's like, do you, can you design the world around you?
Speaker 2: Mm-hmm.
Speaker 0: So you're like, "I need something with this property. Can the system discover and produce that both from, like, a fundamental scientific discovery perspective, but also an industry?" So, so at, like, someone's working in space or defense or semiconductors, and, like, yeah. We're having these issues. We're trying to, uh, achieve, you know, this property of this material or this layer. Can the system accelerate, um, the development of those technologies? So it's, it's, it's very grounded. Um, that's how we'll know it's working.
Speaker 1: It feels like the applications of solving, building an AI physicist, for lack of a better word, that can s- design the real world are so broad. You can apply them to advanced manufacturing. You can apply them to material science, to chemistry, to all, a- anything that inter- a- a- any process where there's R&D with the physical world required, it seems like we'll benefit from breakthroughs that Periodic is working on. Why hasn't it been done before and what is it about this moment in history that makes the right time to attack this problem?
Speaker 0: Maybe one comment is it's difficult.
Speaker 1: What makes it so difficult?
Speaker 0: I mean, I think part of it is the team.
Speaker 2: Mm-hmm.
Speaker 0: So, f- in our view, this has been enabled by frontier technology in the last couple of years. And so Doge and I have been so focused on basically putting together, like, this N-of-one team. Like, these group of physicists, chemists, uh, um, simulation experts, and some of the best machine learning researchers in the world have never been part of one concerted effort. And we feel in order to actually achieve this, you need all these expertise, you need these, these pillars to do this.
Speaker 1: So, when you guys went about designing the team, you know, t- wha- after you left OpenAI and DeepMind, what was the primary heuristic that you, that you used to guide yourself in figuring out who wanted on the team?
Speaker 2: So, in terms of expertise, we wanted to have a LLM expertise covered, a experimental expertise, and simulation. And for each of these, we wanted to have, basically, world-class talent. Um, and of course, for each team, there's actually a lot of sub-teams. It's like a fractal, right? The expertise is very, uh, fractal-like. So for the experimental side, we want to cover solid state chemistry, solid state physics, automation, and kind of the more facilities, like the more, um, operational aspects or experiments. On the simulation side, there's the more kind of theoretical physics parts, there's the more kind of coding aspects of simulations. And the LLM side, of course, there's mid-training, RL, infra. And yeah, for each of these, we tried to get basically the best people who have innovated in these, like, sub-pillars.
Speaker 0: The technology that we think is necessary to do it has really just emerged in the last couple of years.
Speaker 2: Mm-hmm.
Speaker 0: And this data isn't, like, on a Reddit forum or something. Like you, you need to actually go produce experimental data, simulation data. It's siloed across all of these advanced industries. And many of them, while there's a desire, they may not have knowledge of, you know, some of the most recent techniques that's been driving this, this recent wave in AI.
Speaker 1: There was a moment in time when models like... Or, or papers like the GPT-3 paper, for example, that, you know, set language models or few-shot learners and proposed the idea of scaling laws. Uh, and then there was a follow-up paper, if you guys remember, from opening I had that, um, was called, I think, Scaling Laws for Generative Modeling, um, that just showed that as long as you just kept throwing... You scaled up the amount of compute and data in the right combination, um, you could very predictably improve the performance of, of these models. And the theory was that if you just kept doing that, you know, in, in, at, at infinitum, um, there would be a bunch of emergent capabilities. These models would be able to reason about all kinds of problems out of domain, out of distribution. Um, is it... Wouldn't that argue... W- h- how would you, how would you square the circle with that school of thought that, um, that, you know, naively, the current pre-training and post-training, uh, sort of pipelines at most of the frontier labs won't just eventually crack physics as well? Why, why is, is this idea of physical verification, um, so necessary? And are... is that school of, is that school of, sort of reasoning wrong?
Speaker 0: Yeah. Um, excellent question. Uh, scaling laws empirically seem to continue to hold. So that's not in question.
Speaker 1: Mm-hmm.
Speaker 0: But I think there's a question of what is this Y-axis?
Speaker 1: Mm-hmm.
Speaker 0: And that test distribution is very different from, like, what we're talking about. Um, that test distribution, let's say you're pre-training on the internet, might be, you know, a representative set from the internet, and you'll have these sort of predictable scaling properties.... but that's not going to capture that you, you have a very different set of scaling properties with respect to different distributions.
Speaker 1: Mm-hmm.
Speaker 0: So try to make this a little bit more concrete. Let's say hypothetically we're training a coding model, and we have unit tests to provide some reward signal. So the model writes some PR, we check that the unit tests go from failing to passing, and we say, "This was successful, we're gonna reinforce these things." You might say you start optimizing this, and now the system is becoming ever more capable of writing code for its own development. And you have this acceleration, you have this kinda takeoff scenario. Um, code is one of the most promising areas for this, 'cause there's abundant of data online. You have this feedback loop where the system itself can begin to improve itself, and it's, it's a very promising technique, and we're all seeing the benefits of, um, advanced coding models, and it's accelerating quickly. However, that model is not going to then cure cancer. The knowledge simply doesn't exist. It, it doesn't... You need to optimize against the distribution you care about. So that model, while it's gonna be a very valuable tool as a software engineer, it may help a cancer researcher do their analysis, it simply doesn't have the data, the knowledge, or the expertise iterating against that environment.
Speaker 1: Mm-hmm.
Speaker 0: And I think that's just sort of like the fundamental belief we have.
Speaker 2: Yeah, I mean, so actually Liam and I worked on this a bit when we were looking at the scaling laws for, uh, vision models. And, you know, this also came up a lot in the CLIP paper from OpenAI. Like, the in-domain generalization and the out-of-domain generalization are monotonically correlated, but it's not linear necessarily. And so what that means is you can keep improving your model and it'll improve as the power law in domain.
Speaker 1: Mm-hmm.
Speaker 2: And for out-of-domain tasks, by which I mean, as Liam said, the things that you're trying to do that's a bit different than what's in your training set will also improve as a power law, but the slope of the power law may not be good enough. So that you might need to, you know, spend centuries before you get to the result you want. We saw this in the Noam paper, for example. We published a paper where we saw that as you increase the size of your training set, the IID performance, the in-domain performance improves as a power law. Out-of-domain performance also improves as a power law, but depending on what the out-of-domain is, like how far you are from train distribution, that power law might have such a small slope that it's basically useless.
Speaker 1: Mm-hmm.
Speaker 2: Um, so this is one of the reasons we feel like the best way to make progress is to make your target as close to your in-domain training set as possible. And what, the best way of doing this is to basically iterate on changing your training set to be more like what you want to do. Uh, so this is one answer. The other one is actually maybe even simpler. Um, the experimental data we want actually doesn't exist. So for example, if you look at... Like, you wanna, say, learn on the experimental data in literature for synthesis.
Speaker 1: Mm-hmm.
Speaker 2: Turns out the formation enthalpy labels, which is like the energy it takes to basically assemble the atoms in the shape you want, um, is so high that if you train a machine learning model on it, it's not predictive enough to predict the next one. Uh, so, and one of the reasons for this, as Liam mentioned, people don't usually publish negative results, and negative results are usually very context dependent. So what's a negative result for someone might be positive if they do things differently.
Speaker 1: Mm-hmm.
Speaker 2: So, um, yeah, so not only is there this domain shift problem where what you're trying to do might be different than your training set so the power law won't have the large enough slope you want, but the other problem is for some of these things we want to do, there's no data for it. Uh, for example, for superconductivity, there is a lot of data sets you can look at, but the noise floor on them is so high that training on them usually doesn't help.
Speaker 0: Doge, me, the entire team are deep believers in scaling up and scaling laws, but it's just, do a beeline for the thing you care about.
Speaker 1: Mm-hmm.
Speaker 0: Um, and in our case, we care about advancing science, advancing physical R&D. Um, that's, that's sort of like the thesis.
Speaker 1: Is there a tension between being super super bitter lesson pilled and just throwing more compute at the problem and the, I guess, domain specific pipelines that the lab you guys just described will have to focus on? In the case of Periodic, I think you mentioned the first beelines you guys are making are towards super conductivity and magnetism, right? What is it about those domains that make them good candidates for the first, um, for the first few pipelines that Periodic's working on, and why are they just... Are they pit stops along the way to an AI physicist that generalizes across all kinds of domains, or is there a danger of them being essentially off-ramps, um, that don't result in sort of a, a, the, the, the AI sort of scientific super intelligence that, that, that is the North Star for what you guys are doing?
Speaker 2: Yeah, like I feel like, for example, the high temperature superconductivity goal is actually a goal that has so many sub goals in it. It's a bit like when DeepMind and OpenAI started and said, "We're gonna do AGI." But what that meant was they had to do so many things before they got to these cool results. Like for us, if we wanna get a high temperature superconductor, we probably need to get good at autonomous synthesis, autonomous characterization. We need to get good at, um, characterizing different aspects of the material, um, using the LLM to run the simulations correctly. So it's a North Star, and there's so many goals in the, on the way that would be very, I think, impactful for the community.
Speaker 1: Mm-hmm.
Speaker 2: Um, that's one reason. Another reason is I feel like high temperature superconductivity is such a fundamentally interesting question. For example, if you could find a 200 Kelvin superconductor, even before we make any product with it, that in itself says so much about the universe that we didn't know yet. You know, to be able to see such quantum effects on, at such high temperatures I think would be such an update to people's view of how they see the universe. So we feel like it'll be really impactful for humanity even before we make a product out of it. I think that's one of the reasons. A technical reason also is superconductivity is a phase transition, so it's pretty robust to some of these details that we cannot simulate yet. So for example, uh-... that when you make the material, the superconducting temperature usually is more dominated by its kind of crystal fundamental property than, like, defects or microstructure. Whereas there are certain other materials properties where even if the crystal has the property you want, there's so many other factors that you cannot simulate that would prevent you from seeing that property. So superconductivity has this, like, nice, uh, philosophical, uh, upside to it, has this technical upside to it, um, and it like really rallies both the physicists. Like, there are people who studied physics for 40 years and really excited about superconductivity. And there are people who've never studied physics but very excited about superconductivity. It's like quite rare to find a topic that unites the whole team.
Speaker 0: Yeah. I mean, it's... Like Dosh said, it's like we... In order to do this, there are so many foundational pieces to solve. And our tactic is in order to actually get to this goal of AI scientist, you need to make contact, do the full loop somewhere. If you say you're doing this in just like very vague terms, you sort of just end up back on archive papers and textbooks.
Speaker 2: Mm-hmm.
Speaker 0: And so it's, it's really important for us to do the loop, but then create this repeatable process. Like, how do you go from sub-domain to sub-domain? And there's really interesting questions about how well do the ML systems generalize between these things. Um, what is the generalization of a system between, like, superconductivity data to magnetism data, for instance? And maybe that looks very different than its ability to generalize to fluid mechanics.
Speaker 2: Mm-hmm.
Speaker 0: Um, and I think there's, like, fundamental arguments to make there. Um, but the goal is create this repeatable system, prove it, and then just go through the different domains that way.
Speaker 1: So I, I can see the argument for why, um, cracking room temperature superconductivity from an experimental basis is, is extraordinarily valuable for humanity. But you guys are building a startup. And, um, to use an analogy for why you need to have a clear, um, medium-term path or short to medium-term path along the way to a North Star that is both commercially viable and net positive s- to, to society, what we've seen, for example, with other frontier labs that are working on automating white-collar work or, or software knowledge work, is that, you know, th- there's this North Star of an AI researcher. Um, but that along the way, there were a bunch of sub-goals and so on, um, but a concrete kind of application that opened up a ton of commercial value and, and benefits for users on the way to that AI researcher was the idea of, of AI programming, right? Software engineering has become, uh, probably the first major domain, uh, that, that's, that's caused people to really update their priors about h- how, uh, useful AI models are beyond, uh, kind of consumer applications. And in terms of productivity, their impact has been extraordinary just in a few short months. So if the traditional frontier labs as North Star was an AI researcher and the, the path along the way to get there was programming, AI programming, what is that for Periodic?
Speaker 0: Basically, co-pilots for engineers, researchers in advanced industries. So s- maybe perhaps just being in Silicon Valley, we, you know, we really think about like computer-oriented work. Everything is digital, everything is bits. But there's so many industries, like we were kind of talking about a few, like, you know, space, defense, semiconductors, where they're dealing with iteration of materials, of physics, and that's part of their workflow. Like, how are they designing these new technologies, these new devices? And in the absence of data and the absence of good systems, they don't really have particularly good tools.
Speaker 1: Mm-hmm.
Speaker 0: That is our opportunity. And these are massive R&D budgets. So, uh, yeah, while high temp superconductivity is a great North Star, we very much understand that technology and capital are intertwined.
Speaker 1: Mm-hmm.
Speaker 0: Um, we're going to be able to maximally accelerate science if this is a wildly successful commercial entity.
Speaker 1: Mm-hmm.
Speaker 0: And to do so, we want to accelerate advanced manufacturing in all these different industries, become like an intelligence layer in these... for, for all these teams to accelerate their workflow and start reducing their iteration time, get them to better solutions more quickly-
Speaker 1: Mm-hmm
Speaker 0: ... accelerate their researchers and their engineers.
Speaker 1: Let, let's click a little bit deeper on the, that in practice, sort of a day in the life of a Periodic, uh, team member, where, let's say, half the team... Is this roughly right? A- about half the team are ML scientists with machine learning backgrounds, and the remaining half are physical scientists with physics or chemistry backgrounds. Um, how do you start by uniting the cultures, right? How do you take somebody whose primary career so far and work has been experiments in the lab, in, in wet labs, doing physics and chemistry, um, and give them an intuition for ML and vice versa? Because, you know, you guys are both physicists who then had th- the, the, the career trajectory where you also had the chance to be at, at, at Frontier AI labs and were, were part of training systems that are now, uh, considered sort of landmark, hallmark machine learning systems, um, like ChatGPT, like Noam. But for others who might be coming from one domain, how do you get the team to build an intuition for the other?
Speaker 2: Yeah. So this is a great question. I mean, I feel like it's actually crucial for us to make sure these teams work very closely with each other. Um, so one of the things we're seeing is the physics and the chemists need to figure out how to teach the LLM how to reason about these things. Because I think the Frontier AI labs have figured out how to train them on math and logic, but not yet on physics, chemistry. So one thing we're seeing that's been really, I think, productive is the physicists and chemists are thinking about what are the steps we should include in the mid-training, in the RL training, that will teach the LLM how to reason correctly about quantum mechanics, how to reason correctly about these physical systems. Um, another one, of course, is the LLM researchers are learning quite a bit about the physics, the simulation tools, the goals. Uh, so they've been working together really well.... uh, we have weekly teaching sessions where the LLM researchers teach, you know, how the RL loops work, how the data cleaning works. And then the physicists and chemists are teaching about different aspects of the science. Um, the history of science, that's also very important. Uh, so we feel like that's been going really well. And, you know, when we're looking at this is, the things we have to teach the LLM to be able to discover, say, a superconductor, includes being able to read the literature really well. Like, read all the papers, the textbooks, find the relevant parts, and then being able to run simulations, theoretical calculations, and then take action, run experiments. You know, we feel like this is quite similar to the physical R&D researchers in these companies. They have to read the literature, read maybe internal documents or external documents, and then run simulations, run theoretical calculations, and then actually attempt the thing experimentally, learn from that. So we feel like all the progress we're making towards our internal superconductivity or physics goals actually is making our LLMs much better at serving our customers who are doing very similar workflows.
Speaker 1: Yeah, I think just culture, no stupid questions. You can ask just, like, the dumbest, like, physics question, the- the dumbest ML question. And, I mean, there's a few faculty as part of our company and they're actually excellent teachers.
Speaker 2: Mm-hmm.
Speaker 1: Um, so I mean, these, like, learning sessions have been really fantastic. And another thing I noticed is, uh, computer scientists often think in terms of, like, APIs. So scientists will say something and they're always trying to map it. You're like, "Okay, well, what's the input? What's the output? What's the target? How do I map that back?" Whatever, and it's- it's always just like this translation. And I think we also have built up, as part of the team, there's, there's people, like, on these different edges. So like if you have, uh, a simplex of, like, you know, pure ML, LLM, pure experimentalist, pure stimulation, um, there's people who, like, can live in this inside as well.
Speaker 2: Mm-hmm.
Speaker 1: And so they've been, like, excellent bridges for translating between- between these different groups of people. So it's like active learning to, like, learn the other spaces, creating APIs, and then these kinda bridge connector peoples. Um, I think Doge being an excellent example of that.
Speaker 2: Is it a requirement for somebody who wants to join Periodic to have to have an advanced degree in physics or chemistry? Absolutely not. You know, one of the jokes we're making is who was the NBA player who was saying that I'm much closer to LeBron James than you are to me? The, we were saying the opposite of that to candidates, because the amount that even our best physicists doesn't know about physics is much bigger than the amount that they know about physics. So for this new candidate, even if they have no background in physics, how much they have to learn about what we're trying to do is actually not that different on how much the best physicist has to learn. Because there's so much chemistry to learn, so much material science to learn. And I think this is one of the interesting aspects of science today. You know, in the past, in 1800s, there were these physicists that could do so many different things at the frontier. Today, we reached a point where our intellectual knowledge is so large that a leading thinker can usually only advance in one very specific field, and maybe this is actually holding us back. Because say to discover an amazing superconductor, as we keep going back to this example, you have to know so much about chemistry, physics, synthesis, characterization. And unfortunately, I don't think any human knows enough for, about all of these, so we have to collaborate. So I think our team is kind of like a small example of this, where we have, as Liam said, like a lot of different points in that simplex. And for any person, they have so much to learn, but that's true for basically every other scientist. Uh, so for example, I supposedly come from the physics side of it, but I've been learning so much more physics because we now have people from different areas of chemistry in the team, different areas of physics. And I think it's true for LLM researchers as well. I mean, they come in their aspects or LLM that they probably didn't know until they started working with other researchers in our team. So I think it's a great, and it's like a small example of what we're trying to do with the LLM because we're trying to teach this LLM all these different things that we're learning as researchers. It's like a really fun experience, I think. Yeah.
Speaker 1: And what are you finding makes a great researcher at Periodic that's different from what might make a great researcher at OpenAI or Anthropic or DeepMind?
Speaker 2: I would say there's very high overlap. Um, but probably one of the biggest determinants is, "Do you care about this mission?"
Speaker 1: Mm-hmm.
Speaker 2: Is accelerating science, is, to you, is that like the big goal? And I think looking at the tre- the team right now, it's just incredibly mission-driven set of folks who are like, "Yeah, this is the North Star. Let's do that." Um, if someone really wants to improve some mega corps products, yeah, you'd probably be better off at that mega corp in iterating and improving their products. But if you care about scientific discovery, I think Periodic Labs is the best place to do that.
Speaker 1: How big is the team today?
Speaker 2: We're roughly 30, I believe.
Speaker 1: Yeah.
Speaker 2: Awesome.
Speaker 1: And as you think about taking a lot of the research that's going on at the company and deploying that out in the real world, the kinds of customers that we've talked about, space, defense, advanced manufacturing, these are, these are mission-critical industries that are known for being, um, you know, essential to whatever part of the economy they're part of. But often, they're not the most f- they're not the fastest to adopt new technology. Um, how do you think about deploying the kinds of frontier agents that we've talked about that are great at science, great at physics, um, in companies or organizations that might not be anywhere close to as sophisticated as you are, um, in AI or ML?
Speaker 2: Mm-hmm.
Speaker 1: Is- is there... Do you have a working thesis for how to make sure that the arc of progress is not bottlenecked on deployment? Uh, it sounds like you have a fairly good thesis on how to unblock the arc of scientific progress on the research side. But when it comes to deployment, what might be a working theory that you guys are optimistic about that would help get, um, the systems that Periodic is building out into the real world?
Speaker 0: Mm-hmm. Well, maybe one thing that we've noticed in, in our conversations with all these companies is they all are looking for their AI strategy. Um, they understand that, like, the technology's shifting really quickly, and they're looking at how they're doing their work, and it's not changing as quickly as they think it should be.
Speaker 1: Mm-hmm.
Speaker 0: Um, some industries also are losing, like, kinda key expertise, uh, in different fields, and they're, like, losing these, like, senior engineers, senior researchers, and they're like, "Okay, how do we, like, preserve that?" Um, but one thesis is understand, es- you know, kind of thinking about these, like, APIs and thinking about what are the evaluations, what are the biggest bottlenecks for these companies, um, looking at some of the problems they face, and we can map that to our systems, and we say, "Well, we think we can dramatically accelerate this." And so it's not coming in and saying, "Hey, we're gonna transform your fab line on day one. We're gonna transform how you're doing everything. Forget everything."
Speaker 1: Mm-hmm.
Speaker 0: It's like, no, we're gonna, we're gonna solve a really critical problem, well-scoped, very clear evaluations, kind of co-draft that with them, and just show them, like, how powerful this technology can be when you optimize against the thing you care about.
Speaker 1: Mm-hmm.
Speaker 0: Um, so, you know, nothing, uh, particularly, like, uh, surprising here, but you know, sort of like a land-and-expand type method as, as you might expect.
Speaker 1: Mm-hmm.
Speaker 0: But really looking for who are the biggest promoters, um, within that company? What are the biggest problems? Make sure you're solving a very real thing for them, and intersect that with where's our technical capability the highest.
Speaker 1: You know, you were on a call this morning with one of the customers in your pipeline. You don't, we don't need to name who, but what, what were some of the things you heard as their, as their most urgent problems that they'd like for Periodic to solve?
Speaker 2: Uh, so one of them was simulations. You know, they spend a lot of time training people on some of these simulations they need to use, which is critical for their development. Um, and being able to automate those simulations I think would be quite enabling. Uh, the design process, um, and then kind of like some of the small things, like matching the formats, being able to feed, you know, the simulation results into the design pipeline, all of these seem quite important. And then being able to treat the data together in the same place. Uh, what else?
Speaker 0: Well, I think there's a really fundamental question. So, a lot of these companies will rely on retrieval.
Speaker 1: Mm-hmm.
Speaker 0: So, that's sort of like a super lightweight thing. Someone shows up with a neural net, and they're like, "Great, we'll just retrieve over all of your data, and then that's your solution." However, as we've seen with things like ChatGPT and other things, is when you pre-train on the data, when you actually encode the knowledge into the weights, it's not just a retrieval system.
Speaker 1: Mm-hmm.
Speaker 0: You have a richer, deeper understanding of the material. And I think this is a big fundamental challenge. So for instance, for this customer, they can, um, give privileges to their employees and have retrieval as acting on behalf, like, the system acts as the, as the user, and so you can match those same kinda like privileges for access.
Speaker 1: Mm-hmm.
Speaker 0: But if you start doing pre-training or mid-training on different parts, it's like, well, if you pre-train on every piece of data, uh, that might only be accessible to, say, like, the CEO of that company.
Speaker 1: Mm-hmm.
Speaker 0: So then, you have to figure out how do you sort of bucket that knowledge and create different types of systems. Um, but I think right now, like, we're... After talking with the user, they don't seem to have a great solution for sort of distilling all of the knowledge, um, in, into like a single model-
Speaker 1: Mm-hmm
Speaker 0: ... or into a set of models. So they're going, you know, going beyond retrieval to, you know, proper training. And then, I think also the supervised training they're doing is really akin to, like, the early days of ChatGPT where it's like input, output, you have, you know, a few examples. And kinda transforming this new way of thinking was like, no, um, high compute reinforcement learning is really effective.
Speaker 1: Mm-hmm.
Speaker 0: This is how you should think about the strategies it's using. This is how you create effective tool using towards those problems, and this is how you optimize it effectively.
Speaker 1: Could you describe for folks who may not be familiar with it what you mean by mid-training? 'Cause people are familiar with pre-training, they're familiar with post-training, but in the Periodic context, what does mid-training mean?
Speaker 0: Yeah. Sorry for the lingo. So, I, I think this, this term came up, uh, years ago where it's like, well, we had pre-training, we had post-training, but sometimes you need to put in a little bit more knowledge. So, uh, before search worked really well, there was an issue of freshness. So we had pre-trained models, and they have a knowledge cutoff. So there's like a scrape of the internet at that point, but users want more real-time knowledge. So it's like, how do you get that in there? And enter mid-train. Mid-train is basically you're taking new data, new knowledge that's not in the model, and you continue pre-train.
Speaker 1: Mm-hmm.
Speaker 0: And this differs from standard post-training, where post-training typically is more reinforcement learning, supervised, um, learning. And the mechanism is basi- or the, the goal of it is just put a lot of knowledge into the model that doesn't exist before.
Speaker 1: Mm-hmm.
Speaker 0: Um, so that's, that's mid-training in a nutshell.
Speaker 1: And in the Periodic context, does that mean, um, essentially going and injecting a ton of custom sort of data from a- an experimental implementation in a, in a particular customer or particular industry? What is the, what are the sort of the, the lines, the atomic unit, um, that you guys think will, of, of mid-training th- that will improve the capabilities of the models on, on problems that they're just terrible at today?
Speaker 0: I mean, it's just, it's all, all the knowledge. So, it's like you can have very low-level descriptions of, um, physical objects, like-... crystal structures, for instance.
Speaker 1: Mm-hmm.
Speaker 0: And you can also have higher level semantic descriptions of, like, "Well, this is how I made, um, material X, Y, Z."
Speaker 1: Mm-hmm.
Speaker 0: And trying to con- you know, get all this data into the model, um, is really valuable. So it's like simulation data, experimental data, none of this exists. And basically putting that knowledge into the model and making sure that these distributions are connected in some way.
Speaker 1: Mm-hmm.
Speaker 0: And what I mean by that is if you just sort of mix together distribution A, B and C, there's no guarantee of generalization. What you wanna hope to see from these systems is the inclusion of this other dataset is improving performance on the other datasets.
Speaker 1: Mm-hmm.
Speaker 0: And so these are sort of just like, um, machine learning techniques or machine learning problems to solve. Um, but basically, just make it an expert in physics, in chemistry, and where it was deficient before.
Speaker 1: Um, you guys both know that I, I, uh, spent some time running evals on a bunch of these models at the Stanford Physics Lab earlier, earlier this year. And the results were that the models are ter- terrible at scientific analysis.
Speaker 0: 'Cause they weren't trained to do so.
Speaker 1: 'Cause they weren't trained to do so. But on the other hand, you know, many of the, the existing research teams working on the general models are investing in trying to make these better. Is there something about the way you're building periodic that gets the draft off of all of that progress in the base models? Or do you have to start everything from scratch and therefore not be able to be composable with advancements happening in the ba- the mainline models today?
Speaker 2: Yeah, I mean, we benefit from all the different advances. So one of them is the LLMs are getting better. Um, and we definitely benefit from that because we take a pretrained model and then midtrain it, you know, high computer. Another one is the physical simulation tools are getting better. They're open sourcing new ways of simulating, new ways of using machine learning to predict mater- uh, properties. So we get to basically utilize all of those. Um, and it seems like machine learning has made such an impact in the physics and chemistry fields that we expect these improvements to continue.
Speaker 0: I think another thing is, um, when we think about tools for agents, we think of, like, here's a browser, here's a Python, but increasingly, people think about tools as other neural nets, as other agents.
Speaker 1: Right.
Speaker 0: Um, and so if you look at a lot of, like, physics code, it's not particularly deep. It's not... This isn't competition programming. This is, like, kinda like hacky scripts, but you can rely on, uh, some of the best systems for, you know, wherever they spike on. So neural net as a tool to these agents is something that immediately accelerates our work. Um, so you don't have to, like, replicate every- everything.
Speaker 1: There's a historical pattern that a lot of the fundamental research in the physical sciences that, uh, we're talking about here, physics, chemistry, biology, has historically been done at university labs. Um, is there a role at all that the university ecosystem you think will play in periodic's future? Or do you think these are just completely divergent paths?
Speaker 2: Absolutely. I mean, so much of the simulation tooling we use have been developed in academia.
Speaker 1: Mm-hmm.
Speaker 2: Um, s- many of it is in Europe, for example, a lot of the novel synthesis methods. So we definitely benefit from a lot of these different, very deep technical progress. Uh, like for example, all the physical simulation tools are these, you know, complicated Fortran code that in our team, for example, we don't really, like, know how to develop very efficiently. But, um, we feel like there's definitely a very, um, deep connection between academia and industry labs. So for example, recently, a lot of the large scale simulations have been done in industry labs like Microsoft, DeepMind and Meta, but a lot of those tools have been actually developed in academia and then passed on. So there's actually really nice synergy there.
Speaker 0: I think I'd add a few other things too. So like you found when you were evaluating models on their ability to do scientific analysis, they were deficient. This was probably, I mean, not a direct goal for those teams training those models. So I think academia and these collaborations say, "Well, help us un- inform what are the important tasks." Like, how do you do this analysis? What skills do we wanna put in the model? Um, a skill could be a full analysis or a skill could be like a, a smaller primitive as part of a larger analysis. But also secondarily, it's how do you think?
Speaker 1: Mm-hmm.
Speaker 0: So one of the physicists, uh, was looking at the reasoning strategies of one of our models. He's like, "No, this... It's all wrong. It's all wrong." And we're like, "What do you mean?" He's like, "No, this should be thinking higher level. It should be thinking in terms of symmetries."
Speaker 1: Mm-hmm.
Speaker 0: "This is, this is the book, like, that encodes, like, the thinking strategies that will be more effective." And of course, your reinforcement learning environment needs to reward those types of strategies, but given some of the most premier scientists are using these strategies, they're likely effective.
Speaker 1: Mm-hmm.
Speaker 0: And these are types of things where it's like an industry academic partnership can just be so powerful, because industry just simply is blind to these types of analyses, these tools, as well as just this way of thinking.
Speaker 2: Yeah. And there's a way of connecting that to the tool in question as well, because, you know, language is very important, but then in the human brain, we also see other visual processing, like geometric. So it's plausible that while these LLMs will keep, uh, getting better and better, they'll actually benefit from having a geometric reasoning that's separate. So today, we can do that with equivariant graph neural networks. We can do it with diffusion models that are, uh, kinda geometric tools by construction. And the LLM can call them, so then it can have both the language aspect, which is very good for, say, synthesis recipe, but also the geometric aspects, which is very good for representing atoms, uh, just design geometries in general.
Speaker 1: So how are you thinking about deepening periodic's ties with academic labs?
Speaker 2: Yeah, this is very important for us. So we have two major initiatives in this direction. One of them is we're starting an advisory board. Um, this will be kind of-... expertise spanning from superconductivity to solid-state chemistry to physics, and we want to make sure, you know, we're in touch with this kind of long-term research directions. Um, a lot of important government funding goes to these groups, and we wanna have a tight coupling between what's important for them and us. So, uh, this, you know, includes superconductivity expertise, such as ZX Chen from Stanford on the experimental side and Steve Quelson from the theory side. Uh, we also have synthesis expertise on the advisory board from Mercuri Kannatsidis from, uh, Northwestern University, and Chris Wolverton on the high-throughput DFT side. And our second initiative is going to be through a grant program. You know, we really want to enable some of this amazing work going on in academia, and some of their work isn't a good fit for industry. You know, it's best done in academia. So we want to, um, kind of accept grant proposals, and we want to enable and support the kind of work that's gonna help community, especially in relation to LLMs, agents in synthesis, materials discovery, physics modeling. Uh, so, you know, maybe after this show, you can include the link.
Speaker 1: Yeah, we'll include them in the show notes. So for people who might be interested in joining Periodic, what are you guys looking for?
Speaker 0: First off, someone deeply curious. Uh, someone who really wants to understand the machine learning, the science at a deeper level, who wants to make contact with reality, who wants to advance science. Like, this has to be a driving thing. Um, but also pragmatic. Um, what we're trying to do is incredibly challenging, and someone who has, like, very careful process, and they get to... They're solution-oriented. They get to goals quickly. Um, and really, someone world-class along some dimension. Um, we're looking across all these different pillars, so machine learning, experimentalists, simulation. And people who can bring some sort of innovation on what, how do you create a creative ML system? Um, how do you bring new types of tools or new types of thinking to some of these state-of-the-art models? Um, someone who can advance simulations and make it more robust and more reliable with experiment.
Speaker 2: Yeah, and maybe one more thing I'd add is Liam and I have been really looking for a sense of urgency in candidates. Because we want these technologies not in 10 years. You know, we don't want these LMs to start improving science in 10 years. We want them ASAP. So if the candidate feels like a sense of urgency for improving these physical systems, uh, discovering these amazing materials, innovating on superconductivity, they would be a good fit.
Speaker 0: Yeah. If you match all these, please reach out.
Speaker 1: All right. Sounds like we gotta amp up the speed, the scale of stuff happening at Periodic, and we'll put the career links in the show notes. Thanks for coming, guys.