POPULARITY
Editor's note: CuspAI raised a $100m Series A in September and is rumored to have reached a unicorn valuation. They have all-star advisors from Geoff Hinton to Yann Lecun and team of deep domain experts to tackle this next frontier in AI applications.In this episode, Max Welling traces the thread connecting quantum gravity, equivariant neural networks, diffusion models, and climate-focused materials discovery (yes, there is one!!!).We begin with a provocative framing: experiments as computation. Welling describes the idea of a “physics processing unit”—a world in which digital models and physical experiments work together, with nature itself acting as a kind of processor. It's a grounded but ambitious vision of AI for science: not replacing chemists, but accelerating them.Along the way, we discuss:* Why symmetry and equivariance matter in deep learning* The tradeoff between scale and inductive bias* The deep mathematical links between diffusion models and stochastic thermodynamics* Why materials—not software—may be the real bottleneck for AI and the energy transition* What it actually takes to build an AI-driven materials platformMax reflects on moving from curiosity-driven theoretical physics (including work with Gerard ‘t Hooft) toward impact-driven research in climate and energy. The result is a conversation about convergence: physics and machine learning, digital models and laboratory experiments, long-term ambition and incremental progress.Full Video EpisodeTimestamps* 00:00:00 – The Physics Processing Unit (PPU): Nature as the Ultimate Computer* Max introduces the idea of a Physics Processing Unit — using real-world experiments as computation.* 00:00:44 – From Quantum Gravity to AI for Materials* Brandon frames Max's career arc: VAE pioneer → equivariant GNNs → materials startup founder.* 00:01:34 – Curiosity vs Impact: How His Motivation Evolved* Max explains the shift from pure theoretical curiosity to climate-driven impact.* 00:02:43 – Why CaspAI Exists: Technology as Climate Strategy* Politics struggles; technology scales. Why materials innovation became the focus.* 00:03:39 – The Thread: Physics → Symmetry → Machine Learning* How gauge symmetry, group theory, and relativity informed equivariant neural networks.* 00:06:52 – AI for Science Is Exploding (Not Emerging)* The funding surge and why AI-for-Science feels like a new industrial era.* 00:07:53 – Why Now? The Two Catalysts Behind AI for Science* Protein folding, ML force fields, and the tipping point moment.* 00:10:12 – How Engineers Can Enter AI for Science* Practical pathways: curriculum, workshops, cross-disciplinary training.* 00:11:28 – Why Materials Matter More Than Software* The argument that everything—LLMs included—rests on materials innovation.* 00:13:02 – Materials as a Search Engine* The vision: automated exploration of chemical space like querying Google.* 01:14:48 – Inside CuspAI: The Platform Architecture* Generative models + multi-scale digital twin + experiment loop.* 00:21:17 – Automating Chemistry: Human-in-the-Loop First* Start manual → modular tools → agents → increasing autonomy.* 00:25:04 – Moonshots vs Incremental Wins* Balancing lighthouse materials with paid partnerships.* 00:26:22 – Why Breakthroughs Will Still Require Humans* Automation is vertical-specific and iterative.* 00:29:01 – What Is Equivariance (In Plain English)?* Symmetry in neural networks explained with the bottle example.* 00:30:01 – Why Not Just Use Data Augmentation?* The optimization trade-off between inductive bias and data scale.* 00:31:55 – Generative AI Meets Stochastic Thermodynamics* His upcoming book and the unification of diffusion models and physics.* 00:33:44 – When the Book Drops (ICLR?)TranscriptMax: I want to think of it as what I would call a physics processing unit, like a PPU, right? Which is you have digital processing units and then you have physics processing units. So it's basically nature doing computations for you. It's the fastest computer known, as possible even. It's a bit hard to program because you have to do all these experiments. Those are quite bulky, it's like a very large thing you have to do. But in a way it is a computation and that's the way I want to see it. You can do computations in a data center and then you can ask nature to do some computations. Your interface with nature is a bit more complicated. But then these things will have to seamlessly work together to get to a new material that you're interested in.[01:00:44:14 - 01:01:34:08]Brandon: Yeah, it's a pleasure to have Max Woehling as a guest today. Max has done so much over his career that I've been so excited about. If you're in the deep learning community, you probably know Max for his work on variational autocoders, which has literally stood the test of prime or officially stood the test of prime. If you are a scientist, you probably know him for his like, binary work on graph neural networks on equivariance. And if you're a material science, you probably know him about his new startup, CASPAI. Max has a long history doing lots of cool problems. You started in quantum gravity, which is I think very different than all of these other things you worked on. The first question for AI engineers and for scientists, what is the thread in how you think about problems? What is the thread in the type of things which excite you? And how do you decide what is the next big thing you want to work on?[01:01:34:08 - 01:02:41:13]Max: So it has actually evolved a lot. In my young days, let's breathe, I would just follow what I would find super interesting. I have kind of this sensor. I think many people have, but maybe not really sort of use very much, which is like, you get this feeling about getting very excited about some problem. Like it could be, what's inside of a black hole or what's at the boundary of the universe or what are quantum mechanics actually all about. And so I follow that basically throughout my career. But I have to say that as you get older, this changes a little bit in the sense that there's a new dimension coming to it and there's this impact. Going in two-dimensional quantum gravity, you pretty much guaranteed there's going to be no impact on what you do relative, maybe a few papers, but not in this world, this energy scale. As I get closer to retirement, which is fortunately still 10 years away or so, I do want to kind of make a positive impact in the world. And I got pretty worried about climate change.[01:02:43:15 - 01:03:19:11]Max: I think politics seems to have a hard time solving it, especially these days. And so I thought better work on it from the technology side. And that's why we started CaspAI. But there's also a lot of really interesting science problems in material science. And so it's kind of combining both the impact you can make with it as well as the interesting science. So it's sort of these two dimensions, like working on things which you feel there's like, well, there's something very deep going on here. And on the other hand, trying to build tools that can actually make a real impact in the world.[01:03:19:11 - 01:03:39:23]RJ: So the thread that when I look back, look at the different things that you worked out, some of them seem pretty connected, like the physics to equivariance and, yeah, and, uh, gravitational networks, maybe. And that seems to be somewhat related to Casp. Do you have a thread through there?[01:03:39:23 - 01:06:52:16]Max: Yeah. So physics is the thread. So having done, you know, spent a lot of time in theoretical physics, I think there is first very fundamental and exciting questions, like things that haven't actually been figured out in quantum gravity. So that is really the frontier. There's also a lot of mathematical tools that you can use, right? In, for instance, in particle physics, but also in general relativity, sort of symmetry space to play an enormously important role. And this goes all the way to gauge symmetries as well. And so applying these kinds of symmetries to, uh, machine learning was actually, you know, I thought of it as a very deep and interesting mathematical problem. I did this with Taco Cohen and Taco was the main driver behind this, went all the way from just simple, like rotational symmetries all the way to gauge symmetries on spheres and stuff like that. So, and, uh, Maurice Weiler, who's also here, um, when he was a PhD student, he was a very good student with me, you know, he wrote an entire book, which I can really recommend about the role of symmetries in AI and machine learning. So I find this a very deep and interesting problem. So more recently, so I've taken a sort of different path, which is the relationship between diffusion models and that field called stochastic thermodynamics. This is basically the thermodynamics, which is a theory of equilibrium. So but then formulated for out of equilibrium systems. And it turns out that the mathematics that we use for diffusion models, but even for reinforcement learning for Schrodinger bridges for MCMC sampling has the same mathematics as this theoretical, this physical theory of non-equilibrium systems. And that got me very excited. And actually, uh, when I taught a course in, um, Mauschenberg, uh, it is South Africa, close to Cape Town at the African Institute for Mathematical Sciences Ames. And I turned that into a book site. Two years later, the book was finished. I've sent it to the publisher. And this is about the deep relationship between free energy, diffusion models, basically generative AI and stochastic thermodynamics. So it's always some kind of, I don't know, I find physics very deep. I also think a lot about quantum mechanics and it's, it's, it's a completely weird theory that actually nobody really understands. And there's a very interesting story, which is maybe good to tell to connect sort of my PZ back to where I'm now. So I did my PZ with a Nobel Laureate, Gerard the toft. He says the most brilliant man I've ever met. He was never wrong about anything as long as I've seen him. And now he says quantum mechanics is wrong and he has a new theory of quantum mechanics. Nobody understands what he's saying, even though what he's writing down is not mathematically very complex, but he's trying to address this understandability, let's say of quantum mechanics head on. And I find it very courageous and I'm completely fascinated by it. So I'm also trying to think about, okay, can I actually understand quantum mechanics in a more mundane way? So that, you know, without all the weird multiverses and collapses and stuff like that. So the physics is always been the threat and I'm trying to apply the physics to the machine learning to build better algorithms.[01:06:52:16 - 01:07:05:15]Brandon: You are still very involved in understanding and understanding physics and the worlds. Yeah. And just like applications to machine learning or introducing no formalisms. That's really cool.[01:07:05:15 - 01:07:18:02]Max: Yes, I would say I'm not contributing much to physics, but I'm contributing to the interface between physics and science. And that's called AI for science or science or AI is kind of a super, it's actually a new discipline that's emerging.[01:07:18:02 - 01:07:18:19]Speaker 5: Yeah.[01:07:18:19 - 01:07:45:14]Max: And it's not just emerging, it's exploding, I would say. That's the better term because I know you go from investments into like in the hundreds of millions now in the billions. So there's now actually a startup by Jeff Bezos that is at 6.2 billion sheep round. Right. Insane. I guess it's the largest startup ever, I think. And that's in this field, AI for science. It tells you something that we are creating a new bubble here.[01:07:46:15 - 01:07:53:28]Brandon: So why do you think it is? What has changed that has motivated people to start working on AI for science type problems?[01:07:53:28 - 01:08:49:17]Max: So there's two reasons actually. One is that people have been applying sort of the new tools from AI to the sciences, which is quite natural. And there's of course, I think there's two big examples, protein folding is a big one. And the other one is machine learning forest fields or something called machine learning inter-atomic potentials. Both of them have been actually very successful. Both also had something to do with symmetries, which is a little cool. And sort of people in the AI sciences saw an opportunity to apply the tools that they had developed beyond advertised placement, right, or multimedia applications into something that could actually make a very positive impact in society like health, drug development, materials for the energy transition, carbon capture. These are all really cool, impactful applications.[01:08:50:19 - 01:09:42:14]Max: Despite that, the science and the kind of the is also very interesting. I would say the fact that these sort of these two fields are coming together and that we're now at the point that we can actually model these things effectively and move the needle on some of these sort of science sort of methodologies is also a very unique moment, I would say. People recognize that, okay, now we're at the cusp of something new, where it results whether the company is called after. We're at the cusp of something new. And of course that always creates a lot of energy. It's like, okay, there's something, it's like sort of virgin field. It's like nobody's green field. Nobody's been there. I can rush in and I can sort of start harvesting there, right? And I think that's also what's causing a lot of sort of enthusiasm in the fields.[01:09:42:14 - 01:10:12:18]RJ: If you're an AI engineer, basically if the people that listen to this podcast will be in the field, then you maybe don't have a strong science background. How does, but are excited. Most I would say most AI practitioners, BM engineers or scientists would consider themselves scientists and they have some background, a little bit of physics, a little bit of industry college, maybe even graduate school that have been working or are starting out. How does somebody who is not a scientist on a day-to-day basis, how do they get involved?[01:10:12:18 - 01:10:14:28]Max: Well, they can read my book once it's out.[01:10:16:07 - 01:11:05:24]Max: This is basically saying that there is more, we should create curricula that are on this interface. So I'm not sure there is, also we already have some universities actual courses you can take, maybe online courses you can take. These workshops where we are now are actually very good as well. And we should probably have more tutorials before the workshop starts. Actually we've, I've kind of proposed this at some point. It's like maybe first have an hour of a tutorial so that people can get new into the field. There's a lot out there. Most of it is of course inaccessible, but I would say we will create much more books and other contents that is more accessible, including this podcast I would say. So I think it will come. And these days you can watch videos and things. There's a huge amount of content you can go and see.[01:11:05:24 - 01:11:28:28]Brandon: So maybe a follow-up to that. How do people learn and get involved? But why should they get involved? I mean, we have a lot of people who are of our audience will be interested in AI engineering, but they may be looking for bigger impacts in the world. What opportunities does AI for science provide them to make an impact to change the world? That working in this the world of pure bits would not.[01:11:28:28 - 01:11:40:06]Max: So my view is that underlying almost everything is immaterial. So we are focusing a lot on LLMs now, which is kind of the software layer.[01:11:41:06 - 01:11:56:05]Max: I would say if you think very hard, underlying everything is immaterial. So underlying an LLM is a GPU, and underlying a GPU is a wafer on which we will have to deposit materials. Do we want to wait a little bit?[01:12:02:25 - 01:12:11:06]Max: Underlying everything is immaterial. So I was saying, you know, there's the LLM underlying the LLM is a GPU on which it runs. In order to make that GPU,[01:12:12:08 - 01:12:43:20]Max: you have to put materials down on a wafer and sort of shine on it with sort of EUV light in order to etch kind of the structures in. But that's now an actual material problem, because more or less we've reached the limits of scaling things down. And now we are trying to improve further by new materials. So that's a fundamental materials problem. We need to get through the energy transition fast if we don't want to kind of mess up this world. And so there is, for instance, batteries. That's a complete materials problem. There's fuel cells.[01:12:44:23 - 01:13:01:16]Max: There is solar panels. So that they can now make solar panels with new perovskite layers on top of the silicon layers that can capture, you know, theoretically up to 50% of the light, where now we're at, I don't know, maybe 22 or something. So these are huge changes all by material innovation.[01:13:02:21 - 01:13:47:15]Max: And yeah, I think wherever you go, you know, I can probably dig deep enough and then tell you, well, actually, the very foundation of what you're doing is a material problem. And so I think it's just very nice to work on this very, very foundation. And also because I think this is maybe also something that's happening now is we can start to search through this material space. This has never been the case, right? It's like scientists, the normal way of working is you read papers and then you come up with no hypothesis. You do an experiment and you learn, et cetera. So that's a very slow process. Now we can treat this as a search engine. Like we search the internet, we now search the space of all possible molecules, not just the ones that people have made or that they're in the universe, but all of them.[01:13:48:21 - 01:14:42:01]Max: And we can make this kind of fully automated. That's the hope, right? We can just type, it becomes a tool where you type what you want and something starts spinning and some experiments get going. And then, you know, outcome list of materials and then you look at it and say, maybe not. And then you refine your query a little bit. And you kind of do research with this search engine where a huge amount of computation and experimentation is happening, you know, somewhere far away in some lab or some data center or something like this. I find this a very, very promising view of how we can sort of build a much better sort of materials layer underneath almost everything. And also more sustainable materials. Our plastics are polluting the planet. If you come up with a plastic that kind of destroys itself, you know, after, I don't a few weeks, right? And actually becomes a fertilizer. These are things that are not impossible at all. These things can be done, right? And we should do it.[01:14:42:01 - 01:14:47:23]RJ: Can you tell us a little bit just generally about CUSBI and then I have a ton of questions.[01:14:47:23 - 01:14:48:15]Speaker 5: Yeah.[01:14:48:15 - 01:17:49:10]Max: So CUSBI started about 20 months ago and it was because I was worried about I'm still worried about climate change. And so I realized that in order to get, you know, to stay within two degrees, let's say, we would not only have to reduce our emissions to zero by 2050, but then, you know, another half century or even a century of removing carbon dioxide from the atmosphere, not by reducing your emissions, but actually removing it at a rate that's about half the rate that we now emit it. And that is a unsolved problem. But if we don't solve it, two degrees is not going to happen, right? It's going to be much more. And I don't think people quite understand how bad that can be, like four degrees, like very bad. So this technology needs to be developed. And so this was my and my co-founder, Chet Edwards, motivation to start this startup. And also because, you know, we saw the technology was ready, which is also very good. So if you're, you know, the time is right to do it. And yeah, so we now in the meanwhile, we've grown to about 40 people. We've kind of collected 130 million investment into the company, which is for a European company is quite a lot. I would say it's interesting that right after that, you know, other startups got even more. So that's kind of tells you how fast this is growing. But yeah, we are we are now at the we've built the platform, of course, but it's for a series of material classes and it needs to be constantly expanded to new material classes. And it can be more automated because, you know, we know putting LLMs in as the whole thing gets more and more automated. And now we're moving to sort of high throughput experimentation. So connecting the actual platform, which is computational, to the experiments so that you can get also get fast feedback from experiments. And I kind of think of experiments as something you do at the end, although that's what we've been doing so far. I want to think of it as what I would call a sort of a physics processing unit, like a PPU, right, which is you have digital processing units and then you have physics processing units. So it's basically nature doing computations for you. It's the fastest computer known as possible, even. It's a bit hard to program because you have to do all these experiments. Those are quite, quite bulky. It's like a very large thing you have to do. But in a way, it is a computation. And that's the way I want to see it. So I want to you can do computations in a data center and then you can ask nature to do some computations. Your interface with nature is a bit more complicated. But then these things will have to seamlessly work together to get to a new material that you're interested in. And that's the vision we have. We don't say super intelligence because I don't quite know what it means and I don't want to oversell it. But I do want to automate this process and give a very powerful tool in the hands of the chemists and the material scientists.[01:17:49:10 - 01:18:01:02]Brandon: That actually brings up a question I wanted to ask you. First of all, can you talk about your platform to like whatever degree, like explain kind of how it works and like what you your thought processes was in developing it?[01:18:01:02 - 01:20:47:22]Max: Yeah, I think it's been surprisingly, it's not rocket science, I would say. It's not rocket science in the sense of the design and basically the design that, you know, I wrote down at the very beginning. It's still more or less the design, although you add things like I wasn't thinking very much about multi-scale models and as the common are rated that actually multi-scale is very important. And the beginning, I wasn't thinking very much about self-driving labs. But now I think, you know, we are now at the stage we should be adding that. And so there is sort of bits and details that we're adding. But more or less, it's what you see in the slide decks here as well, which is there is a generative component that you have to train to generate candidates. And then there is a digital twin, multi-scale, multi-fidelity digital twin, which you walk through the steps of the ladder, you know, they do the cheap things first, you weed out everything that's obviously unuseful, and then you go to more and more expensive things later. And so you narrow things down to a small number. Those go into an experiment, you know, do the experiment, get feedback, etc. Now, things that also have been more recently added is sort of more agentic sort of parts. You know, we have agents that search the literature and come up with, you know, actually the chemical literature and come up with, you know, chemical suggestions for doing experiments. We have agents which sort of autonomously orchestrate all of the computations and the experiments that need to be done. You know, they're in various stages of maturity and they can be continuously improved, I would say. And so that's basically I don't think that part. There's rocket science, but, you know, the design of that thing is not like surprising. What is it's surprising hard to actually build it. Right. So that's that's the thing that is where the moat is in the data that you can get your hands on and the and actually building the platform. And I would say there's two people in particular I want to call out, which is Felix Hunker, who is actually, you know, building the scientific part of the platform and Sandra de Maria, who is building the sort of the skate that is kind of this the MLOps part of the platform. Yeah. And so and recently we also added sort of Aaron Walsh to our team, who is a very accomplished scientist from Imperial College. We're very happy about that. He's going to be a chief science officer. And we also have a partnerships team that sort of seeks out all the customers because I think this is one thing I find very important. In print, it's so complex to do to actually bring a material to the real world that you must do this, you know, in collaboration with sort of the domain experts, which are the companies typically. So we always we only start to invest in the direction if we find a good industrial partner to go on that journey with us.[01:20:47:22 - 01:20:55:12]Brandon: Makes a lot of sense. Over the evolution of the platform, did you find that you that human intervention, human,[01:20:56:18 - 01:21:17:01]Brandon: I guess you could start out with a pure, you could imagine two directions when you start up making everything purely automatic, automated, agentic, so on. And then later on, you like find that you need to have more human input and feedback different steps. Or maybe did you start out with having human feedback? You have lots of steps and then like kind of, yeah, figure out ways to remove, you know,[01:21:17:01 - 01:22:39:18]Max: that is the second one. So you build tools for you. So it's much more modular than you think. But it's like, we need these tools for this application. We need these tools. So you build all these tools, and then you go through a workflow actually in the beginning just manually. So you put them in a first this tool, then run this to them or this with sithery. So you put them in a workflow and then you figure out, oh, actually, you know, this this porous material that we are trying to make actually collapses if you shake it a bit. Okay, then you add a new tool that says test for stability. Right. Yeah. And so there's more and more tools. And then you build the agent, which could be a Bayesian optimizer, or it could be an actual other them, you know, maybe trained to be a good chemist that will then start to use all these tools in the right way in the right order. Yeah. Right. But in the beginning, it's like you as a chemist are putting the workflow together. And then you think about, okay, how am I going to automate this? Right. For one very easy question you can ask yourself is, you know, every time somebody who is not a super expert in DFT, yeah, and he wants to do a calculation has to go to somebody who knows DFT. And so could you start to automate that away, which is like, okay, make it so user friendly, so that you actually do the right DFT for the right problem and for the right length of time, and you can actually assess whether it's a good outcome, etc. So you start to automate smaller small pieces and bigger pieces, etc. And in the end, the whole thing is automated.[01:22:39:18 - 01:22:53:25]Brandon: So your philosophy is you want to provide a set of specific tools that make it so that the scientists making decisions are better informed and less so trying to create an automated process.[01:22:53:25 - 01:23:22:01]Max: I think it's this is sort of the same where you're saying because, yes, we want to automate, yeah, but we don't see something very soon where the chemists and the domain expert is out of the loop. Yeah, but it but it's a retreat, right? It's like, okay, so first, you need an expert to tell you precisely how to set the parameters of the DFT calculation. Okay, maybe we can take that out. We can maybe automate that, right? And so increasingly, more of these things are going to be removed.[01:23:22:01 - 01:23:22:19]Speaker 5: Yeah.[01:23:22:19 - 01:24:33:25]Max: In the end, the vision is it will be a search engine where you where somebody, a chemist will type things and we'll get candidates, but the chemist will still decide what is a good material and what is not a good material out of that list, right? And so the vision of a completely dark lab, where you can close the door and you just say, just, you know, find something interesting and then it will it will just figure out what's interesting and we'll figure out, you know, it's like, oh, I found this new material to blah, blah, blah, blah, right? That's not the vision I have. He's not for, you know, a long time. So for me, it's really empowering the domain experts that are sitting in the companies and in universities to be much faster in developing their materials. And I should say, it's also good to be a little humble at times, because it is very complicated, you know, to bring it to make it and to bring it into the real world. And there are people that are doing this for the entire lives. Yeah. Right. And it's like, I wonder if they scratch their head and say, well, you know, how are you going to completely automate that away, like in the next five years? I don't think that's going to happen at all.[01:24:35:01 - 01:24:39:24]Max: Yeah. So to me, it's an increasingly powerful tool in the hands of the chemists.[01:24:39:24 - 01:25:04:02]RJ: I have a question. You've talked before about getting people interested based on having, you know, sort of a big breakthrough in materials, incremental change. I'm curious what you think about the platform you have now in are sort of stepping towards and how are you chasing the big change or is this like incremental or is there they're not mutually exclusive, obviously, but what do you think about that?[01:25:04:02 - 01:26:04:27]Max: We follow a mixed strategy. So we are definitely going after a big material. Again, we do this with a partner. I'm not going to disclose precisely what it is, but we have our own kind of long term goal. You could call it lighthouse or, you know, sort of moonshot or whatever, but it is going to be a really impactful material that we want to develop as a proof point that it can be done and that it will make it into the into the real world and that AI was essential in actually making it happen. At the same time, we also are quite happy to work with companies that have more modest goals. Like I would say one is a very deep partnership where you go on a journey with a company and that's a long term commitment together. And the other one is like somebody says, I knew I need a force field. Can you help me train this force field and then maybe analyze this particular problem for me? And I'll pay you a bunch of money for that. And then maybe after that we'll see. And that's fine too. Right. But we prefer, you know, the deep partnerships where we can really change something for the good.[01:26:04:27 - 01:26:22:02]RJ: Yeah. And do you feel like from a platform standpoint you're ready for that or what are the things that and again, not asking you to disclose proprietary secret sauce, but what are the things generally speaking that need to happen from where we are to where to get those big breakthroughs?[01:26:22:02 - 01:28:40:01]Max: What I find interesting about this field is that every time you build something, it's actually immediately useful. Right. And so unlike quantum computing, which or nuclear fusion, so you work for 20, 30, 40 years and nothing, nothing, nothing, nothing. And then it has to happen. Right. And when it happens, it's huge. So it's quite different here because every time you introduce, so you go to a customer and you say, so what do you need? Right. So we work, let's say, on a problem like a water filtration. We want to remove PFAS from water. Right. So we do this with a company, Camira. So they are a deep partner for us. Right. So we on a journey together. I think that the breakthrough will happen with a lot of human in the loop because there is the chemists who have a whole lot more knowledge of their field and it's us who will help them with training, having a new message. And in that kind of interface, these interactions, something beautiful will happen and that will have to happen first before this field will really take off, I think. And so in the sense that it's not a bubble, let's put it that way. So that's people see that as actual real what's happening. So in the beginning, it will be very, you know, with a lot of humans in the loop, I would say, and I would I would hope we will have this new sort of breakthrough material before, you know, everything is completely automated because that will take a while. And also it is very vertical specific. So it's like completely automating something for problem A, you know, you can probably achieve it, but then you'll sort of have to start over again for problem B because, you know, your experimental setup looks very different in the machines that you characterize your materials look very different. Even the models in your platform will have to be retrained and fine tuned to the new class. So every time, you know, you have a lot of learnings to transfer, but also, you know, the problems are actually different. And so, yes, I would want that breakthrough material before it's completely automated, which I think is kind of a long term vision. And I would say every time you move to something new, you'll have to start retraining and humans will have to come in again and say, okay, so what does this problem look like? And now sort of, you know, point the the machine again, you know, in the new direction and then and then use it again.[01:28:40:01 - 01:28:47:17]RJ: For the non-scientists among us, me included a bit of a scientist. There's a lot of terminology. You mentioned DFT,[01:28:49:00 - 01:29:01:11]RJ: you equivariance we've talked about. Can you sort of explain in engineering terms or the level of sophistication and engineering? Well, how what is equivariance?[01:29:01:11 - 01:29:55:01]Max: So equivariance is the infusion of symmetry in neural networks. So if I build a neural network, let's say that needs to recognize this bottle, right, and then I rotate the bottle, it will then actually have to completely start again because it has no idea that the rotated bottle. Well, actually, the input that represents a rotated bottle is actually rotated bottle. It just doesn't understand that. Right. If you build equivariance in basically once you've trained it in one orientation, it will understand it in any other orientation. So that means you need a lot less data to train these models. And these are constraints on the weights of the model. So so basically you have to constrain the way such data to understand it. And you can build it in, you can hard code it in. And yeah, this the symmetry groups can be, you know, translations, rotations, but also permutations. I can graph neural network, their permutations and then physics, of course, as many more of these groups.[01:29:55:01 - 01:30:01:08]RJ: To pray devil's advocate, why not just use data augmentation by your bottle is in all the different orientations?[01:30:01:08 - 01:30:58:23]Max: As an option, it's just not exact. It's like, why would you go through the work of doing all that? Where you would really need an infinite number of augmentations to get it completely right. Where you can also hard code it in. Now, I have to say sometimes actually data augmentation works even better than hard coding the equivariance in. And this is something to do with the fact that if you constrain the optimization, the weights before the optimization starts, the optimization surface or objective becomes more complicated. And so it's harder to find good minima. So there is also a complicated interplay, I think, between the optimization process and these constraints you put in your network. And so, yeah, you'll hear kind of contradicting claims in this field. Like some people and for certain applications, it works just better than not doing it. And sometimes you hear other people, if you have a lot of data and you can do data augmentation, then actually it's easier to optimize them and it actually works better than putting the equivariance in.[01:30:58:23 - 01:31:07:16]Brandon: Do you think there's kind of a bitter lesson for mathematically founded models and strategies for doing deep learning?[01:31:07:16 - 01:31:46:06]Max: Yeah, ultimately it's a trade-off between data and inductive bias. So if your inductive bias is not perfectly correct, you have to be careful because you put a ceiling to what you can do. But if you know the symmetry is there, it's hard to imagine there isn't a way to actually leverage it. But yeah, so there is a bitter lesson. And one of the bitter lessons is you should always make sure your architecture is scale, unless you have a tiny data set, in which case it doesn't matter. But if you, you know, the same bitter lessons or lessons that you can draw in LLM space are eventually going to be true in this space as well, I think.[01:31:47:10 - 01:31:55:01]RJ: Can you talk a little bit about your upcoming book and tell the listeners, like, what's exciting about it? Yeah, I should read it.[01:31:55:01 - 01:33:42:20]Max: So this book is about, it's called Generative AI and Stochastic Thermodynamics. It basically lays bare the fact that the mathematics that goes into both generative AI, which is the technology to generate images and videos, and this field of non-equilibrium statistical mechanics, which are systems of molecules that are just moving around and relaxing to the ground state, or that you can control to have certain, you know, be in a certain state, the mathematics of these two is actually identical. And so that's fascinating. And in fact, what's interesting is that Jeff Hinton and Radford Neal already wrote down the variational free energy for machine learning a long time ago. And there's also Carl Friston's work on free energy principle and active entrance. But now we've related it to this very new field in physics, which is called stochastic thermodynamics or non-equilibrium thermodynamics, which has its own very interesting theorems, like fluctuation theorems, which we don't typically talk about, but we can learn a lot from. And I think it's just it can sort of now start to cross fertilize. When we see that these things are actually the same, we can, like we did for symmetries, we can now look at this new theory that's out there, developed by these very smart physicists, and say, okay, what can we take from here that will make our algorithms better? At the same time, we can use our models to now help the scientists do better science. And so it becomes a beautiful cross-fertilization between these two fields. The book is rather technical, I would say. And it takes all sorts of things that have been done as stochastic thermodynamics, and all sorts of models that have been done in the machine learning literature, and it basically equates them to each other. And I think hopefully that sense of unification will be revealing to people.[01:33:42:20 - 01:33:44:05]RJ: Wait, and when is it out?[01:33:44:05 - 01:33:56:09]Max: Well, it depends on the publisher now. But I hope in April, I'm going to give a keynote at ICLR. And it would be very nice if they have this book in my hand. But you know, it's hard to control these kind of timelines.[01:33:56:09 - 01:33:58:19]RJ: Yeah, I'm looking forward to it. Great.[01:33:58:19 - 01:33:59:25]Max: Thank you very much. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.latent.space/subscribe
【お知らせ】11月から「SDGsを話そう」は原則、水〜金曜に配信を固め、「MEDIA TALK」は週の前半を中心に配信します。「ニュースの現場から」は変わらず毎日配信です。 【番組内容】陸軍最大の特攻出撃基地があった鹿児島・知覧の特攻平和会館。近年、見学者や研修ツアーが多く訪れるといいます。「国のため命を捨てた」特攻隊員の手紙が来場者の心を震わせる一方で、出撃命令書を書く役割だった100歳の男性に当時の話を聞くと、複雑な記憶が浮かんできました。 ※2025年10月20日に収録しました。前後編の前編です。後編は11月6日に配信します。 【今回取り上げた記者サロン、関連記事】記者サロン「戦後80年、その先へ」 https://ciy.digital.asahi.com/ciy/11017395 特攻隊員は「志願」だったのか 死へ送り出す命令書、書く手が震えた https://www.asahi.com/articles/AST8B2123T8BUTIL03LM.html?iref=omny 官僚だった知事は消息不明になった 沖縄戦の「英雄」は虚像か実像か https://www.asahi.com/articles/AST885DWCT88UTIL038M.html?iref=omny ★記者サロンのDiscordを開設しましたオンラインの座談会をDiscordで随時開催します。こちらからコミュニティにご参加くださいhttps://bit.ly/kishasalon_discord ★配信中・受付中の記者サロン一覧はこちら!https://www.asahi.com/eventcalendar/ 【出演者】平川仁(東京社会部)MC 飯島啓史安仁周(MC・音源編集) 【おねがい】朝日新聞ポッドキャストは、みなさまからの購読料で配信しています。番組継続のため、会員登録をお願いします! https://t.asahi.com/womz 【朝ポキ情報】アプリで記者と対話 https://t.asahi.com/won1 交流はdiscord https://bit.ly/asapoki_discord おたよりフォーム https://bit.ly/asapoki_otayori 朝ポキTV https://www.youtube.com/@asapoki_official メルマガ https://bit.ly/asapoki_newsletter 広告ご検討の企業様は https://t.asahi.com/asapokiguide 番組検索ツール https://bit.ly/asapoki_cast 最新情報はX https://bit.ly/asapoki_twitter 番組カレンダー https://bit.ly/asapki_calendar 全話あります公式サイト https://bit.ly/asapoki_lp See omnystudio.com/listener for privacy information.
Sign up for Alex's first live cohort, about Hierarchical Model building!Get 25% off "Building AI Applications for Data Scientists and Software Engineers"Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work!Visit our Patreon page to unlock exclusive Bayesian swag ;)Takeaways:Why GPs still matter: Gaussian Processes remain a go-to for function estimation, active learning, and experimental design – especially when calibrated uncertainty is non-negotiable.Scaling GP inference: Variational methods with inducing points (as in GPflow) make GPs practical on larger datasets without throwing away principled Bayes.MCMC in practice: Clever parameterizations and gradient-based samplers tighten mixing and efficiency; use MCMC when you need gold-standard posteriors.Bayesian deep learning, pragmatically: Stochastic-gradient training and approximate posteriors bring Bayesian ideas to neural networks at scale.Uncertainty that ships: Monte Carlo dropout and related tricks provide fast, usable uncertainty – even if they're approximations.Model complexity ≠ model quality: Understanding capacity, priors, and inductive bias is key to getting trustworthy predictions.Deep Gaussian Processes: Layered GPs offer flexibility for complex functions, with clear trade-offs in interpretability and compute.Generative models through a Bayesian lens: GANs and friends benefit from explicit priors and uncertainty – useful for safety and downstream decisions.Tooling that matters: Frameworks like GPflow lower the friction from idea to implementation, encouraging reproducible, well-tested modeling.Where we're headed: The future of ML is uncertainty-aware by default – integrating UQ tightly into optimization, design, and deployment.Chapters:08:44 Function Estimation and Bayesian Deep Learning10:41 Understanding Deep Gaussian Processes25:17 Choosing Between Deep GPs and Neural Networks32:01 Interpretability and Practical Tools for GPs43:52 Variational Methods in Gaussian Processes54:44 Deep Neural Networks and Bayesian Inference01:06:13 The Future of Bayesian Deep Learning01:12:28 Advice for Aspiring Researchers
How does a star form? How does the universe form? And how can we use every bit of astronomical data to answer those questions? To find out, Dr. Charles Liu and co-host Allen Liu welcome astrostatistician Sabrina Berger, all the way from Melbourne, Australia, where she's currently pursuing her PhD. As always, though, we start off with the day's joyfully cool cosmic thing, the new radioastronomy photographs of Callisto, one of the moons of Jupiter, taken by ALMA. Sabrina talks about her own low-frequency radio astronomy research looking for hydrogen in the very early reionization period of the universe when the first galaxies were forming. (Be warned: we dive into the difficulties ionization poses for trying to discern these early processes, including a side trip into quantum mechanics, the hyperfine transition of neutral hydrogen at 21cm depicted on the plaque attached to the Pioneer spacecraft, and even the Cosmic Background Radiation.) You'll also hear how Sabrina is innovatively using GPS satellites to help calibrate large radioastronomy telescope arrays. For our first student question, Derek asks, “I heard that black holes can form right after the Big Bang, before stars do. How is that possible?” Sabrina describes these primordial black holes, and, although none have been confirmed yet, that there have been a number of papers published recently on the subject. In fact, one paper suggesting that the as-yet-undiscovered “Planet 9” could even be one of these primordial black holes. And then, finally, we get to the subject of astrostatistics, Sabrina's area of expertise. She explains that it allows you to harness every piece of information that you're observing in astronomy and to answer questions like “How does a star form?” or “How does the universe form?” You'll hear about huge data sets, the use of artificial intelligence, field level inferences… and the MCMC, or the Markov chain Monte Carlo used in statistics. (If you don't know what that is, you're not alone, and our own resident mathematician Allen helps Sabrina untangle the complexity with a cotton ball analogy that blew Chuck and Sabrina's collective minds!) For our next student question, Wally asks, “Why is redshift one like nine billion years ago, bur redshift two only two billion years before that, and redshift three only one billion years before that?” As Chuck says, “that's a little complicated,” just before he, Allen and Sabrina proceed to explain how we measure universal expansion, the passage of time, and the “stretching” of light. Our next conversation is one of the most controversial we've ever had and revolves around who Sabrina thinks makes the best espresso, Australia, Italy or a “Third Wave Coffee Shop” like we have here in the US. You'll hear about why there's an ISSpresso machine on the ISS – and how the Italian Space Agency invented a way to make an espresso in zero-g! Plus, you'll hear a little about the work-life balance in Australia and how wonderful astronomy down under is. (Check out our Patreon for the story behind the Australian Aboriginal "Emu-in-the-sky" constellation.) If you'd like to know more about Sabrina, you can find her on Twitter and Blue Sky @sabrinastronomy or check out her research on her website. We hope you enjoy this episode of The LIUniverse, and, if you do, please support us on Patreon. Credits for Images Used in this Episode: An image of Jupiter's icy moon Callisto, photographed by NASA's Galileo spacecraft in 2001. – Credit: NASA/Galileo Photograph of Jupiter taken in 2019. The four fainter objects are four of its moons (left to right): Callisto, Ganymede, Io, and Europa. – Credit: Creative Commons / Rehman Abubakr ALMA images of Callisto – Credit: Maria Camarca et al 2025 Planet. Sci. J. 6 183. See the ALMA/Callisto paper: “A Multifrequency Global View of Callisto's Thermal Properties from ALMA”: https://iopscience.iop.org/article/10.3847/PSJ/ade7ee Timeline of the universe. – Credit: NASA, ESA, CSA, STScI The Pioneer plaques, attached to the Pioneer 10 and 11 spacecraft. – Credit: NASA Sedna orbit with solar system (Sun, Jupiter, Saturn, Uranus, Neptune and Pluto visible) and positions on Jan 1, 2017 – Credit: Creative Commons / Tom Ruen Redshift and universe expansion. As light travels from great distances to Hubble's mirrors, it is stretched to longer and longer red wavelengths, or cosmologically redshifted, as the universe expands. – Credit: NASA, ESA, Leah Hustak (STScI) The ISSpresso machine on the International Space Station.– Credit: NASA Astronaut Samantha Cristoforetti drinking espresso out of the cup on ISS, 2015 – Credit: NASA #liuniverse #charlesliu #allenliu #sciencepodcast #astronomypodcast #sabrinaberger #astrostatistician #astrostatistics #redshift #blackholes #primordialblackholes #callisto #alma #planet9 #sedna #universeexpansion #isspresso
Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!Intro to Bayes Course (first 2 lessons free)Advanced Regression Course (first 2 lessons free)Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work!Visit our Patreon page to unlock exclusive Bayesian swag ;)Takeaways:INLA is a fast, deterministic method for Bayesian inference.INLA is particularly useful for large datasets and complex models.The R INLA package is widely used for implementing INLA methodology.INLA has been applied in various fields, including epidemiology and air quality control.Computational challenges in INLA are minimal compared to MCMC methods.The Smart Gradient method enhances the efficiency of INLA.INLA can handle various likelihoods, not just Gaussian.SPDs allow for more efficient computations in spatial modeling.The new INLA methodology scales better for large datasets, especially in medical imaging.Priors in Bayesian models can significantly impact the results and should be chosen carefully.Penalized complexity priors (PC priors) help prevent overfitting in models.Understanding the underlying mathematics of priors is crucial for effective modeling.The integration of GPUs in computational methods is a key future direction for INLA.The development of new sparse solvers is essential for handling larger models efficiently.Chapters:06:06 Understanding INLA: A Comparison with MCMC08:46 Applications of INLA in Real-World Scenarios11:58 Latent Gaussian Models and Their Importance15:12 Impactful Applications of INLA in Health and Environment18:09 Computational Challenges and Solutions in INLA21:06 Stochastic Partial Differential Equations in Spatial Modeling23:55 Future Directions and Innovations in INLA39:51 Exploring Stochastic Differential Equations43:02 Advancements in INLA Methodology50:40 Getting Started with INLA56:25 Understanding Priors in Bayesian ModelsThank you to my Patrons for making this episode possible!Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad
Headlines: Supreme Court Challenge: Legal challenge against the WA Government's approval of the Woodside Gas Project.NT Government: Abandonment of climate emissions targets.Iran Conflict:Human Rights Activists in Iran report at least 639 Iranians killed and 1,329 injured since June 13 due to Israeli attacks.Iran's retaliatory strikes result in 24 Israeli deaths.NSW Anti-Protest Law:New laws grant police arbitrary powers and restrict protests near religious buildings, linked to antisemitic incidents and criminal conspiracies, not peaceful demonstrations.NSW police powers report || hereVivien Langford from the Climate Action (Monday 5pm on 3cr) show brings us to the streets of Sydney with 8 mins from out side the supreme court in Macquarie street yesterday about how NSW Minns Government rushed through laws giving police powers they did not need or want at the time of the Dynamite Caravan story was being built up as part of a wave of antisemitism.Voices 4 Palestine || hereAPAN's Nasser Mashni (Palestine Remembered on 3cr Saturday mornings 9:30am) speech at Melbourne Free Palestine rally on Sunday the 15th of June.Song - Qlowksy - SurrenderNICA Proteus || hereWe have a chat with Katherine Cawthorn about the remarkable new show that NICA is presenting Proteus and learn what goes on behind the scenes. National Institute of Circus Arts Proteus 18-28 June at their Green street theatre Prahran.This is the Week || hereComrade Kevin recaps a huge week in news with his trademark razor sharp satire. MCMC CEO Bernadette Thomas interview || hereMerri Creek Management Committee (MCMC) CEO Bernadette Thomas joins us to talk about all things Merri Creek and why this very special committee is something that everyone should know about and support, doing vital work all up and down the Merri and getting the community involved as much as they can. They are currently also fundraising so if you have a few spare dollars be sure to send them that way!Song - Maysa Daw - Live Free
「オルトフォン、“新定番MCカートリッジ”「MC Xシリーズ」。ステンレス一体成型フレーム&新磁気回路採用」 オルトフォンジャパンは、MCカートリッジ新シリーズ“Xシリーズ”を5月より発売する。
In this episode Gudrun speaks with Nadja Klein and Moussa Kassem Sbeyti who work at the Scientific Computing Center (SCC) at KIT in Karlsruhe. Since August 2024, Nadja has been professor at KIT leading the research group Methods for Big Data (MBD) there. She is an Emmy Noether Research Group Leader, and a member of AcademiaNet, and Die Junge Akademie, among others. In 2025, Nadja was awarded the Committee of Presidents of Statistical Societies (COPSS) Emerging Leader Award (ELA). The COPSS ELA recognizes early career statistical scientists who show evidence of and potential for leadership and who will help shape and strengthen the field. She finished her doctoral studies in Mathematics at the Universität Göttingen before conducting a postdoc at the University of Melbourne as a Feodor-Lynen fellow by the Alexander von Humboldt Foundation. Afterwards she was a Professor for Statistics and Data Science at the Humboldt-Universität zu Berlin before joining KIT. Moussa joined Nadja's lab as an associated member in 2023 and later as a postdoctoral researcher in 2024. He pursued a PhD at the TU Berlin while working as an AI Research Scientist at the Continental AI Lab in Berlin. His research primarily focuses on deep learning, developing uncertainty-based automated labeling methods for 2D object detection in autonomous driving. Prior to this, Moussa earned his M.Sc. in Mechatronics Engineering from the TU Darmstadt in 2021. The research of Nadja and Moussa is at the intersection of statistics and machine learning. In Nadja's MBD Lab the research spans theoretical analysis, method development and real-world applications. One of their key focuses is Bayesian methods, which allow to incorporate prior knowledge, quantify uncertainties, and bring insights to the “black boxes” of machine learning. By fusing the precision and reliability of Bayesian statistics with the adaptability of machine and deep learning, these methods aim to leverage the best of both worlds. The KIT offers a strong research environment, making it an ideal place to continue their work. They bring new expertise that can be leveraged in various applications and on the other hand Helmholtz offers a great platform in that respect to explore new application areas. For example Moussa decided to join the group at KIT as part of the Helmholtz Pilot Program Core-Informatics at KIT (KiKIT), which is an initiative focused on advancing fundamental research in informatics within the Helmholtz Association. Vision models typically depend on large volumes of labeled data, but collecting and labeling this data is both expensive and prone to errors. During his PhD, his research centered on data-efficient learning using uncertainty-based automated labeling techniques. That means estimating and using the uncertainty of models to select the helpful data samples to train the models to label the rest themselves. Now, within KiKIT, his work has evolved to include knowledge-based approaches in multi-task models, eg. detection and depth estimation — with the broader goal of enabling the development and deployment of reliable, accurate vision systems in real-world applications. Statistics and data science are fascinating fields, offering a wide variety of methods and applications that constantly lead to new insights. Within this domain, Bayesian methods are especially compelling, as they enable the quantification of uncertainty and the incorporation of prior knowledge. These capabilities contribute to making machine learning models more data-efficient, interpretable, and robust, which are essential qualities in safety-critical domains such as autonomous driving and personalized medicine. Nadja is also enthusiastic about the interdisciplinarity of the subject — repeatedly changing the focus from mathematics to economics to statistics to computer science. The combination of theoretical fundamentals and practical applications makes statistics an agile and important field of research in data science. From a deep learning perspective, the focus is on making models both more efficient and more reliable when dealing with large-scale data and complex dependencies. One way to do this is by reducing the need for extensive labeled data. They also work on developing self-aware models that can recognize when they're unsure and even reject their own predictions when necessary. Additionally, they explore model pruning techniques to improve computational efficiency, and specialize in Bayesian deep learning, allowing machine learning models to better handle uncertainty and complex dependencies. Beyond the methods themselves, they also contribute by publishing datasets that help push the development of next-generation, state-of-the-art models. The learning methods are applied across different domains such as object detection, depth estimation, semantic segmentation, and trajectory prediction — especially in the context of autonomous driving and agricultural applications. As deep learning technologies continue to evolve, they're also expanding into new application areas such as medical imaging. Unlike traditional deep learning, Bayesian deep learning provides uncertainty estimates alongside predictions, allowing for more principled decision-making and reducing catastrophic failures in safety-critical application. It has had a growing impact in several real-world domains where uncertainty really matters. Bayesian learning incorporates prior knowledge and updates beliefs as new data comes in, rather than relying purely on data-driven optimization. In healthcare, for example, Bayesian models help quantify uncertainty in medical diagnoses, which supports more risk-aware treatment decisions and can ultimately lead to better patient outcomes. In autonomous vehicles, Bayesian models play a key role in improving safety. By recognizing when the system is uncertain, they help capture edge cases more effectively, reduce false positives and negatives in object detection, and navigate complex, dynamic environments — like bad weather or unexpected road conditions — more reliably. In finance, Bayesian deep learning enhances both risk assessment and fraud detection by allowing the system to assess how confident it is in its predictions. That added layer of information supports more informed decision-making and helps reduce costly errors. Across all these areas, the key advantage is the ability to move beyond just accuracy and incorporate trust and reliability into AI systems. Bayesian methods are traditionally more expensive, but modern approximations (e.g., variational inference or last layer inference) make them feasible. Computational costs depend on the problem — sometimes Bayesian models require fewer data points to achieve better performance. The trade-off is between interpretability and computational efficiency, but hardware improvements are helping bridge this gap. Their research on uncertainty-based automated labeling is designed to make models not just safer and more reliable, but also more efficient. By reducing the need for extensive manual labeling, one improves the overall quality of the dataset while cutting down on human effort and potential labeling errors. Importantly, by selecting informative samples, the model learns from better data — which means it can reach higher performance with fewer training examples. This leads to faster training and better generalization without sacrificing accuracy. They also focus on developing lightweight uncertainty estimation techniques that are computationally efficient, so these benefits don't come with heavy resource demands. In short, this approach helps build models that are more robust, more adaptive to new data, and significantly more efficient to train and deploy — which is critical for real-world systems where both accuracy and speed matter. Statisticians and deep learning researchers often use distinct methodologies, vocabulary and frameworks, making communication and collaboration challenging. Unfortunately, there is a lack of Interdisciplinary education: Traditional academic programs rarely integrate both fields. It is necessary to foster joint programs, workshops, and cross-disciplinary training can help bridge this gap. From Moussa's experience coming through an industrial PhD, he has seen how many industry settings tend to prioritize short-term gains — favoring quick wins in deep learning over deeper, more fundamental improvements. To overcome this, we need to build long-term research partnerships between academia and industry — ones that allow for foundational work to evolve alongside practical applications. That kind of collaboration can drive more sustainable, impactful innovation in the long run, something we do at methods for big data. Looking ahead, one of the major directions for deep learning in the next five to ten years is the shift toward trustworthy AI. We're already seeing growing attention on making models more explainable, fair, and robust — especially as AI systems are being deployed in critical areas like healthcare, mobility, and finance. The group also expect to see more hybrid models — combining deep learning with Bayesian methods, physics-based models, or symbolic reasoning. These approaches can help bridge the gap between raw performance and interpretability, and often lead to more data-efficient solutions. Another big trend is the rise of uncertainty-aware AI. As AI moves into more high-risk, real-world applications, it becomes essential that systems understand and communicate their own confidence. This is where uncertainty modeling will play a key role — helping to make AI not just more powerful, but also more safe and reliable. The lecture "Advanced Bayesian Data Analysis" covers fundamental concepts in Bayesian statistics, including parametric and non-parametric regression, computational techniques such as MCMC and variational inference, and Bayesian priors for handling high-dimensional data. Additionally, the lecturers offer a Research Seminar on Selected Topics in Statistical Learning and Data Science. The workgroup offers a variety of Master's thesis topics at the intersection of statistics and deep learning, focusing on Bayesian modeling, uncertainty quantification, and high-dimensional methods. Current topics include predictive information criteria for Bayesian models and uncertainty quantification in deep learning. Topics span theoretical, methodological, computational and applied projects. Students interested in rigorous theoretical and applied research are encouraged to explore our available projects and contact us for further details. The general advice of Nadja and Moussa for everybody interested to enter the field is: "Develop a strong foundation in statistical and mathematical principles, rather than focusing solely on the latest trends. Gain expertise in both theory and practical applications, as real-world impact requires a balance of both. Be open to interdisciplinary collaboration. Some of the most exciting and meaningful innovations happen at the intersection of fields — whether that's statistics and deep learning, or AI and domain-specific areas like medicine or mobility. So don't be afraid to step outside your comfort zone, ask questions across disciplines, and look for ways to connect different perspectives. That's often where real breakthroughs happen. With every new challenge comes an opportunity to innovate, and that's what keeps this work exciting. We're always pushing for more robust, efficient, and trustworthy AI. And we're also growing — so if you're a motivated researcher interested in this space, we'd love to hear from you." Literature and further information Webpage of the group G. Nuti, Lluis A.J. Rugama, A.-I. Cross: Efficient Bayesian Decision Tree Algorithm, arxiv Jan 2019 Wikipedia: Expected value of sample information C. Howson & P. Urbach: Scientific Reasoning: The Bayesian Approach (3rd ed.). Open Court Publishing Company. ISBN 978-0-8126-9578-6, 2005. A.Gelman e.a.: Bayesian Data Analysis Third Edition. Chapman and Hall/CRC. ISBN 978-1-4398-4095-5, 2013. Yu, Angela: Introduction to Bayesian Decision Theory cogsci.ucsd.edu, 2013. Devin Soni: Introduction to Bayesian Networks, 2015. G. Nuti, L. Rugama, A.-I. Cross: Efficient Bayesian Decision Tree Algorithm, arXiv:1901.03214 stat.ML, 2019. M. Carlan, T. Kneib and N. Klein: Bayesian conditional transformation models, Journal of the American Statistical Association, 119(546):1360-1373, 2024. N. Klein: Distributional regression for data analysis , Annual Review of Statistics and Its Application, 11:321-346, 2024 C.Hoffmann and N.Klein: Marginally calibrated response distributions for end-to-end learning in autonomous driving, Annals of Applied Statistics, 17(2):1740-1763, 2023 Kassem Sbeyti, M., Karg, M., Wirth, C., Klein, N., & Albayrak, S. (2024, September). Cost-Sensitive Uncertainty-Based Failure Recognition for Object Detection. In Uncertainty in Artificial Intelligence (pp. 1890-1900). PMLR. M. K. Sbeyti, N. Klein, A. Nowzad, F. Sivrikaya and S. Albayrak: Building Blocks for Robust and Effective Semi-Supervised Real-World Object Detection pdf. To appear in Transactions on Machine Learning Research, 2025 Podcasts Learning, Teaching, and Building in the Age of AI Ep 42 of Vanishing Gradient, Jan 2025. O. Beige, G. Thäter: Risikoentscheidungsprozesse, Gespräch im Modellansatz Podcast, Folge 193, Fakultät für Mathematik, Karlsruher Institut für Technologie (KIT), 2019.
In this episode Gudrun speaks with Nadja Klein and Moussa Kassem Sbeyti who work at the Scientific Computing Center (SCC) at KIT in Karlsruhe. Since August 2024, Nadja has been professor at KIT leading the research group Methods for Big Data (MBD) there. She is an Emmy Noether Research Group Leader, and a member of AcademiaNet, and Die Junge Akademie, among others. In 2025, Nadja was awarded the Committee of Presidents of Statistical Societies (COPSS) Emerging Leader Award (ELA). The COPSS ELA recognizes early career statistical scientists who show evidence of and potential for leadership and who will help shape and strengthen the field. She finished her doctoral studies in Mathematics at the Universität Göttingen before conducting a postdoc at the University of Melbourne as a Feodor-Lynen fellow by the Alexander von Humboldt Foundation. Afterwards she was a Professor for Statistics and Data Science at the Humboldt-Universität zu Berlin before joining KIT. Moussa joined Nadja's lab as an associated member in 2023 and later as a postdoctoral researcher in 2024. He pursued a PhD at the TU Berlin while working as an AI Research Scientist at the Continental AI Lab in Berlin. His research primarily focuses on deep learning, developing uncertainty-based automated labeling methods for 2D object detection in autonomous driving. Prior to this, Moussa earned his M.Sc. in Mechatronics Engineering from the TU Darmstadt in 2021. The research of Nadja and Moussa is at the intersection of statistics and machine learning. In Nadja's MBD Lab the research spans theoretical analysis, method development and real-world applications. One of their key focuses is Bayesian methods, which allow to incorporate prior knowledge, quantify uncertainties, and bring insights to the “black boxes” of machine learning. By fusing the precision and reliability of Bayesian statistics with the adaptability of machine and deep learning, these methods aim to leverage the best of both worlds. The KIT offers a strong research environment, making it an ideal place to continue their work. They bring new expertise that can be leveraged in various applications and on the other hand Helmholtz offers a great platform in that respect to explore new application areas. For example Moussa decided to join the group at KIT as part of the Helmholtz Pilot Program Core-Informatics at KIT (KiKIT), which is an initiative focused on advancing fundamental research in informatics within the Helmholtz Association. Vision models typically depend on large volumes of labeled data, but collecting and labeling this data is both expensive and prone to errors. During his PhD, his research centered on data-efficient learning using uncertainty-based automated labeling techniques. That means estimating and using the uncertainty of models to select the helpful data samples to train the models to label the rest themselves. Now, within KiKIT, his work has evolved to include knowledge-based approaches in multi-task models, eg. detection and depth estimation — with the broader goal of enabling the development and deployment of reliable, accurate vision systems in real-world applications. Statistics and data science are fascinating fields, offering a wide variety of methods and applications that constantly lead to new insights. Within this domain, Bayesian methods are especially compelling, as they enable the quantification of uncertainty and the incorporation of prior knowledge. These capabilities contribute to making machine learning models more data-efficient, interpretable, and robust, which are essential qualities in safety-critical domains such as autonomous driving and personalized medicine. Nadja is also enthusiastic about the interdisciplinarity of the subject — repeatedly changing the focus from mathematics to economics to statistics to computer science. The combination of theoretical fundamentals and practical applications makes statistics an agile and important field of research in data science. From a deep learning perspective, the focus is on making models both more efficient and more reliable when dealing with large-scale data and complex dependencies. One way to do this is by reducing the need for extensive labeled data. They also work on developing self-aware models that can recognize when they're unsure and even reject their own predictions when necessary. Additionally, they explore model pruning techniques to improve computational efficiency, and specialize in Bayesian deep learning, allowing machine learning models to better handle uncertainty and complex dependencies. Beyond the methods themselves, they also contribute by publishing datasets that help push the development of next-generation, state-of-the-art models. The learning methods are applied across different domains such as object detection, depth estimation, semantic segmentation, and trajectory prediction — especially in the context of autonomous driving and agricultural applications. As deep learning technologies continue to evolve, they're also expanding into new application areas such as medical imaging. Unlike traditional deep learning, Bayesian deep learning provides uncertainty estimates alongside predictions, allowing for more principled decision-making and reducing catastrophic failures in safety-critical application. It has had a growing impact in several real-world domains where uncertainty really matters. Bayesian learning incorporates prior knowledge and updates beliefs as new data comes in, rather than relying purely on data-driven optimization. In healthcare, for example, Bayesian models help quantify uncertainty in medical diagnoses, which supports more risk-aware treatment decisions and can ultimately lead to better patient outcomes. In autonomous vehicles, Bayesian models play a key role in improving safety. By recognizing when the system is uncertain, they help capture edge cases more effectively, reduce false positives and negatives in object detection, and navigate complex, dynamic environments — like bad weather or unexpected road conditions — more reliably. In finance, Bayesian deep learning enhances both risk assessment and fraud detection by allowing the system to assess how confident it is in its predictions. That added layer of information supports more informed decision-making and helps reduce costly errors. Across all these areas, the key advantage is the ability to move beyond just accuracy and incorporate trust and reliability into AI systems. Bayesian methods are traditionally more expensive, but modern approximations (e.g., variational inference or last layer inference) make them feasible. Computational costs depend on the problem — sometimes Bayesian models require fewer data points to achieve better performance. The trade-off is between interpretability and computational efficiency, but hardware improvements are helping bridge this gap. Their research on uncertainty-based automated labeling is designed to make models not just safer and more reliable, but also more efficient. By reducing the need for extensive manual labeling, one improves the overall quality of the dataset while cutting down on human effort and potential labeling errors. Importantly, by selecting informative samples, the model learns from better data — which means it can reach higher performance with fewer training examples. This leads to faster training and better generalization without sacrificing accuracy. They also focus on developing lightweight uncertainty estimation techniques that are computationally efficient, so these benefits don't come with heavy resource demands. In short, this approach helps build models that are more robust, more adaptive to new data, and significantly more efficient to train and deploy — which is critical for real-world systems where both accuracy and speed matter. Statisticians and deep learning researchers often use distinct methodologies, vocabulary and frameworks, making communication and collaboration challenging. Unfortunately, there is a lack of Interdisciplinary education: Traditional academic programs rarely integrate both fields. It is necessary to foster joint programs, workshops, and cross-disciplinary training can help bridge this gap. From Moussa's experience coming through an industrial PhD, he has seen how many industry settings tend to prioritize short-term gains — favoring quick wins in deep learning over deeper, more fundamental improvements. To overcome this, we need to build long-term research partnerships between academia and industry — ones that allow for foundational work to evolve alongside practical applications. That kind of collaboration can drive more sustainable, impactful innovation in the long run, something we do at methods for big data. Looking ahead, one of the major directions for deep learning in the next five to ten years is the shift toward trustworthy AI. We're already seeing growing attention on making models more explainable, fair, and robust — especially as AI systems are being deployed in critical areas like healthcare, mobility, and finance. The group also expect to see more hybrid models — combining deep learning with Bayesian methods, physics-based models, or symbolic reasoning. These approaches can help bridge the gap between raw performance and interpretability, and often lead to more data-efficient solutions. Another big trend is the rise of uncertainty-aware AI. As AI moves into more high-risk, real-world applications, it becomes essential that systems understand and communicate their own confidence. This is where uncertainty modeling will play a key role — helping to make AI not just more powerful, but also more safe and reliable. The lecture "Advanced Bayesian Data Analysis" covers fundamental concepts in Bayesian statistics, including parametric and non-parametric regression, computational techniques such as MCMC and variational inference, and Bayesian priors for handling high-dimensional data. Additionally, the lecturers offer a Research Seminar on Selected Topics in Statistical Learning and Data Science. The workgroup offers a variety of Master's thesis topics at the intersection of statistics and deep learning, focusing on Bayesian modeling, uncertainty quantification, and high-dimensional methods. Current topics include predictive information criteria for Bayesian models and uncertainty quantification in deep learning. Topics span theoretical, methodological, computational and applied projects. Students interested in rigorous theoretical and applied research are encouraged to explore our available projects and contact us for further details. The general advice of Nadja and Moussa for everybody interested to enter the field is: "Develop a strong foundation in statistical and mathematical principles, rather than focusing solely on the latest trends. Gain expertise in both theory and practical applications, as real-world impact requires a balance of both. Be open to interdisciplinary collaboration. Some of the most exciting and meaningful innovations happen at the intersection of fields — whether that's statistics and deep learning, or AI and domain-specific areas like medicine or mobility. So don't be afraid to step outside your comfort zone, ask questions across disciplines, and look for ways to connect different perspectives. That's often where real breakthroughs happen. With every new challenge comes an opportunity to innovate, and that's what keeps this work exciting. We're always pushing for more robust, efficient, and trustworthy AI. And we're also growing — so if you're a motivated researcher interested in this space, we'd love to hear from you." Literature and further information Webpage of the group G. Nuti, Lluis A.J. Rugama, A.-I. Cross: Efficient Bayesian Decision Tree Algorithm, arxiv Jan 2019 Wikipedia: Expected value of sample information C. Howson & P. Urbach: Scientific Reasoning: The Bayesian Approach (3rd ed.). Open Court Publishing Company. ISBN 978-0-8126-9578-6, 2005. A.Gelman e.a.: Bayesian Data Analysis Third Edition. Chapman and Hall/CRC. ISBN 978-1-4398-4095-5, 2013. Yu, Angela: Introduction to Bayesian Decision Theory cogsci.ucsd.edu, 2013. Devin Soni: Introduction to Bayesian Networks, 2015. G. Nuti, L. Rugama, A.-I. Cross: Efficient Bayesian Decision Tree Algorithm, arXiv:1901.03214 stat.ML, 2019. M. Carlan, T. Kneib and N. Klein: Bayesian conditional transformation models, Journal of the American Statistical Association, 119(546):1360-1373, 2024. N. Klein: Distributional regression for data analysis , Annual Review of Statistics and Its Application, 11:321-346, 2024 C.Hoffmann and N.Klein: Marginally calibrated response distributions for end-to-end learning in autonomous driving, Annals of Applied Statistics, 17(2):1740-1763, 2023 Kassem Sbeyti, M., Karg, M., Wirth, C., Klein, N., & Albayrak, S. (2024, September). Cost-Sensitive Uncertainty-Based Failure Recognition for Object Detection. In Uncertainty in Artificial Intelligence (pp. 1890-1900). PMLR. M. K. Sbeyti, N. Klein, A. Nowzad, F. Sivrikaya and S. Albayrak: Building Blocks for Robust and Effective Semi-Supervised Real-World Object Detection pdf. To appear in Transactions on Machine Learning Research, 2025 Podcasts Learning, Teaching, and Building in the Age of AI Ep 42 of Vanishing Gradient, Jan 2025. O. Beige, G. Thäter: Risikoentscheidungsprozesse, Gespräch im Modellansatz Podcast, Folge 193, Fakultät für Mathematik, Karlsruher Institut für Technologie (KIT), 2019.
SOMECITY THE FINAL 2024-2025。今年も心が躍りまくりの3日間でした。配信MCと敗者復活戦MCとして話させてもらった3人による振り返りトーク✏️JIZO・BigMo・VIBESへの質問、お悩み相談、ポッドキャストで話して欲しい事の提案、は こちらから!
第2次世界大戦では、婦選活動を率いた市川房枝は「戦争協力」の道を選ばざるを得ませんでした。ドイツでは第1次世界大戦で多くの同性愛者が志願兵として戦いましたが、ナチスは迫害しました。戦時中の女性や同性愛者の経験から、今、くむべき教訓を考えます。*2025年2月7日に収録しました。 【関連記事】市川房枝が迫られた「三択」 ひたむきな活動は戦争協力に変質したhttps://www.asahi.com/articles/ASSDV0FJNSDVUPQJ00CM.html?iref=omny 「誰かのために」利用され、切り捨てられた 戦争がさらす国家の素顔https://www.asahi.com/articles/ASSDV1075SDVUSPT00DM.html?iref=omny 連載「100年をたどる旅」音声で追体験 取材班が舞台裏を語りますhttps://www.asahi.com/articles/AST1P4FWQT1PDIFI006M.html?iref=omny 【出演・スタッフ】高重治香(論説委員)花房吾早子(大阪社会部)MC 岸上渉MC・音源編集 南日慶子 【おねがい】朝日新聞ポッドキャストは、みなさまからの購読料で配信しています。番組継続のため、会員登録をお願いします! http://t.asahi.com/womz 【朝ポキ情報】アプリで記者と対話 http://t.asahi.com/won1 交流はdiscord https://bit.ly/asapoki_discord おたよりフォーム https://bit.ly/asapoki_otayori 朝ポキTV https://bit.ly/asapoki_youtube_ メルマガ https://bit.ly/asapoki_newsletter 広告ご検討の企業様は http://t.asahi.com/asapokiguide 番組検索ツール https://bit.ly/asapoki_cast 最新情報はX https://bit.ly/asapoki_twitter 番組カレンダー https://bit.ly/asapki_calendar 全話あります公式サイト https://bit.ly/asapoki_lp See omnystudio.com/listener for privacy information.
Jaguars players OL Tyler Shatley, TE Evan Engram and LB Foye Oluokun speak with the media on Wednesday ahead of Week 13 vs. Texans. OL Tyler Shatley: 00:00 - 02:55 TE Evan Engram: 02:56 - 09:10 LB Foye Oluokun: 09:11 - 14:29See omnystudio.com/listener for privacy information.
Jaguars QB Trevor Lawrence meets with the media on Wednesday of Week 13 ahead of the matchup against the Houston Texans.See omnystudio.com/listener for privacy information.
Jaguars players OL Tyler Shatley, TE Evan Engram and LB Foye Oluokun speak with the media on Wednesday ahead of Week 13 vs. Texans. OL Tyler Shatley: 00:00 - 02:55 TE Evan Engram: 02:56 - 09:10 LB Foye Oluokun: 09:11 - 14:29See omnystudio.com/listener for privacy information.
Jaguars QB Trevor Lawrence meets with the media on Wednesday of Week 13 ahead of the matchup against the Houston Texans.See omnystudio.com/listener for privacy information.
The award of the second 5G spectrum has taken many by surprise when U Mobile, the smallest of telco players won it. However, this has raised more questions as to the reasons, technical specifications and also the benefit to the country for this award that the regulator, MCMC has yet to answer. Professor Dr Ong Kian Ming, Pro Vice Chancellor for External Engagement, at Taylor's University gives us his perspective and the implications of this on Malaysia Inc.Image Credit: shutterstock.com
这期太难写简介了, 本期节目仅代表578两位主播个人言论 节目内容如果冒犯到各位,我们先道歉哈 本期主播: MC鲍 MC黄
In this week's episode Greg tries to ambush Patrick by bringing back the popular feature Pop Quiz, this time with a statistical acronym theme, only to pretty much get crushed by Patrick in the end. Along they way they also discuss: Wow That's Fantastic, QR codes and octogenarians, Questionable Rectum, catharsis, grassy knolls, petards, Sean ringtones, pity minutes, apologies to Roy Levy, bad clock management, asteroid Roombas, pitching beach balls, statistical sock puppets, and the DIC talk. Stay in contact with Quantitude! Web page: quantitudepod.org TwitterX: @quantitudepod YouTube: @quantitudepod Merch: redbubble.com
本期节目有大量“迅”有关的话题 如果你身边有孩子的话,请戴上您的耳机 让我们一起和我台前线记者感受一下阿姆斯特丹的夜色吧 ps:本期和艺术毫不相关,请听友老师们放心 本期主播: MC鲍 MC黄
阿姆斯特丹真是一个适合周五飞过去一趟看看艺术品的城市 这么近那么美,周末到阿姆斯特丹看看这两幅有故事的画吧 本期主播: MC鲍 MC黄
Audio Siar Keluar Sekejap Episod 122 antaranya membincangkan penghalaan sistem nama domain (DNS) yang ingin dikuatkuasakan oleh MCMC dan kewajipan pensijilan halal yang sedang dikaji oleh JAKIM. Keluar Sekejap juga menyentuh strategi geopolitik PMX, Anwar Ibrahim susulan lawatan ke Rusia baru-baru ini. Rujukan : https://www.economist.com/asia/2024/08/29/why-does-the-west-back-the-wrong-asian-leaders https://www.scmp.com/week-asia/economics/article/3276605/malaysias-china-plus-one-gold-rush-stumbles-over-us-tariff-threat?module=top_story&pgtype=section Dikesempatan ini, Keluar Sekejap ingin mengucapkan terima kasih kepada Viu yang telah menaja episod kali ini. Viu, sebuah perkhidmatan penstriman video over-the-top (OTT) membawakan kandungan dan hiburan Asia terbaik termasuk drama, filem and program gaya hidup. Nikmati siri terbaru Viu Original The Secret setiap Khamis hanya di Viu. Tebus kod: KSSECRET sebelum 15 October di laman web / mobile web browser www.viu.com untuk dapatkan akses 30 hari percuma. Ikuti viu di Facebook : https://www.facebook.com/ViuMalaysia?mibextid=LQQJ4d Instagram : https://www.instagram.com/viumalaysia?igsh=NzZhdXExeG5qeGhh Youtube : https://youtube.com/@viumalaysia?si=qGVMM5SvNsy9aDPR X : https://x.com/viu_my?s=21&t=SZk0EUfxxKqTZWrHYJdi6w Tiktok : https://www.tiktok.com/@viu.malaysia?_t=8pb7W52LRM8&_r=1
Last week, news emerged that the Malaysian Communications and Multimedia Commission (MCMC) directed all Internet service providers to implement public DNS (Domain Name System) redirection for businesses, enterprises, and government agencies by September 30. Although Communications Minister Fahmi Fadzil defended the decision, he has asked MCMC to put the plan on hold pending stakeholder engagement. We discuss the rationale and implication of the policy with Alex Wong of SoyaCincau.Image Credit: shutterstock.com
童润中是你我, 童润中也不是你我 阶级斗争这四个字分量是不是有点大? 过好日子是不是会更重要点呢? 但是真理确实也需要捍卫。 你有什么想法请留言告诉我们 本期主播: MC鲍 MC黄
怎么说呢,套用一句王建国老师的梗 We Are Back,我们是自行车 欢迎各位微博搜索578广播或者评论区留言加入【578广播花好月圆俱乐部听友群】 本期主播: MC鲍 MC黄
朝ポキと記者サロンで同じようにMCを務めると、視聴者に意図しない印象を与えてしまう…。記者サロンMCの大野由衣は就任当初、そう悩んだといいます。いま意識していることは?反省を次につなげる姿勢に、安仁周と江向彩也夏も感銘を受けました。 ※2024年8月21日に収録しました。前後編の後編です。前編は《スマホ1台で録音したあの日 初代キャップ・大野由衣が語る朝ポキの黎明期(前編)#700》です。 【関連コンテンツ】連載「まなび場天声人語」 https://www.asahi.com/rensai/list.html?id=801岸田氏本人が話す「戦略ミス」 総裁選、はしご外されてhttps://www.asahi.com/articles/ASNBF34QCNBCUEHF007.html?iref=omny砂上の国家 満洲のスパイ戦 - プレミアムAhttps://www.asahi.com/special/manchukuo-spying/?iref=omny 【記者サロンの一覧はこちら】記者イベントカレンダーhttps://www.asahi.com/eventcalendar/?iref=omny 【出演者】大野由衣安仁周江向彩也夏(MC・編集) 【朝ポキ情報】ご感想はおたよりフォーム → https://bit.ly/asapoki_otayori番組カレンダー→ https://bit.ly/asapki_calendar出演者名検索ツール→ https://bit.ly/asapoki_cast最新情報はX(旧ツイッター)→ https://bit.ly/asapoki_twitter交流はコミュニティ → https://bit.ly/asapoki_communityテロップ付きはYouTube → https://bit.ly/asapoki_youtube_こぼれ話はメルマガ → https://bit.ly/asapoki_newsletter全話あります公式サイト → https://bit.ly/asapoki_lp広告ご検討の企業様は → http://t.asahi.com/asapokiguideメールはこちら → podcast@asahi.comSee omnystudio.com/listener for privacy information.
Earlier this, the Asia Internet Coalition (AIC), an industry group, sent a letter to the Prime Minister calling for a halt to new social media licensing. The Malaysian Communications and Multimedia Commission (MCMC) has since questioned AIC's representation of tech companies, noting discrepancies and Grab Malaysia's claim of not being consulted. We turn to Derek John Fernandez, a Commission Member of the MCMC, for insights on these allegations.Image Credit: Malaysia Communications and Multimedia Commission (MCMC)
Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!My Intuitive Bayes Online Courses1:1 Mentorship with meOur theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work!Visit our Patreon page to unlock exclusive Bayesian swag ;)Takeaways:Use mini-batch methods to efficiently process large datasets within Bayesian frameworks in enterprise AI applications.Apply approximate inference techniques, like stochastic gradient MCMC and Laplace approximation, to optimize Bayesian analysis in practical settings.Explore thermodynamic computing to significantly speed up Bayesian computations, enhancing model efficiency and scalability.Leverage the Posteriors python package for flexible and integrated Bayesian analysis in modern machine learning workflows.Overcome challenges in Bayesian inference by simplifying complex concepts for non-expert audiences, ensuring the practical application of statistical models.Address the intricacies of model assumptions and communicate effectively to non-technical stakeholders to enhance decision-making processes.Chapters:00:00 Introduction to Large-Scale Machine Learning11:26 Scalable and Flexible Bayesian Inference with Posteriors25:56 The Role of Temperature in Bayesian Models32:30 Stochastic Gradient MCMC for Large Datasets36:12 Introducing Posteriors: Bayesian Inference in Machine Learning41:22 Uncertainty Quantification and Improved Predictions52:05 Supporting New Algorithms and Arbitrary Likelihoods59:16 Thermodynamic Computing01:06:22 Decoupling Model Specification, Data Generation, and InferenceThank you to my Patrons for making this episode possible!Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, Tim Gasser, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Calculating Natural Latents via Resampling, published by johnswentworth on June 6, 2024 on The AI Alignment Forum. So you've read some of our previous natural latents posts, and you're sold on the value proposition. But there's some big foundational questions still unanswered. For example: how do we find these natural latents in some model, if we don't know in advance what they are? Examples in previous posts conceptually involved picking some latents out of the ether (like e.g. the bias of a die), and then verifying the naturality of that latent. This post is about one way to calculate natural latents, in principle, when we don't already know what they are. The basic idea is to resample all the variables once simultaneously, conditional on the others, like a step in an MCMC algorithm. The resampled variables turn out to be a competitively optimal approximate natural latent over the original variables (as we'll prove in the post). Toward the end, we'll use this technique to calculate an approximate natural latent for a normal distribution, and quantify the approximations. The proofs will use the graphical notation introduced in Some Rules For An Algebra Of Bayes Nets. Some Conceptual Foundations What Are We Even Computing? First things first: what even is "a latent", and what does it even mean to "calculate a natural latent"? If we had a function to "calculate natural latents", what would its inputs be, and what would its outputs be? The way we use the term, any conditional distribution (λ,xP[Λ=λ|X=x]) defines a "latent" variable Λ over the "observables" X, given the distribution P[X]. Together P[X] and P[Λ|X] specify the full joint distribution P[Λ,X]. We typically think of the latent variable as some unobservable-to-the-agent "generator" of the observables, but a latent can be defined by any extension of the distribution over X to a distribution over Λ and X. Natural latents are latents which (approximately) satisfy some specific conditions, namely that the distribution P[X,Λ] (approximately) factors over these Bayes nets: Intuitively, the first says that Λ mediates between the Xi's, and the second says that any one Xi gives approximately the same information about Λ as all of X. (This is a stronger redundancy condition than we used in previous posts; we'll talk about that change below.) So, a function which "calculates natural latents" takes in some representation of a distribution (xP[X]) over "observables", and spits out some representation of a conditional distribution (λ,xP[Λ=λ|X=x]), such that the joint distribution (approximately) factors over the Bayes nets above. For example, in the last section of this post, we'll compute a natural latent for a normal distribution. The function to compute that latent: Takes in a covariance matrix ΣXX for X, representing a zero-mean normal distribution P[X]. Spits out a covariance matrix ΣΛΛ for Λ and a cross-covariance matrix ΣΛX, together representing the conditional distribution of a latent Λ which is jointly zero-mean normal with X. … and the joint normal distribution over Λ,X represented by those covariance matrices approximately factors according to the Bayes nets above. Why Do We Want That, Again? Our previous posts talk more about the motivation, but briefly: two different agents could use two different models with totally different internal (i.e. latent) variables to represent the same predictive distribution P[X]. Insofar as they both use natural latents, there's a correspondence between their internal variables - two latents over the same P[X] which both approximately satisfy the naturality conditions must contain approximately the same information about X. So, insofar as the two agents both use natural latents internally, we have reason to expect that the internal latents of one can be faithfully translated int...
Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!My Intuitive Bayes Online Courses1:1 Mentorship with meHow does the world of statistical physics intertwine with machine learning, and what groundbreaking insights can this fusion bring to the field of artificial intelligence?In this episode, we delve into these intriguing questions with Marylou Gabrié. an assistant professor at CMAP, Ecole Polytechnique in Paris. Having completed her PhD in physics at École Normale Supérieure, Marylou ventured to New York City for a joint postdoctoral appointment at New York University's Center for Data Science and the Flatiron's Center for Computational Mathematics.As you'll hear, her research is not just about theoretical exploration; it also extends to the practical adaptation of machine learning techniques in scientific contexts, particularly where data is scarce.In this conversation, we'll traverse the landscape of Marylou's research, discussing her recent publications and her innovative approaches to machine learning challenges, latest MCMC advances, and ML-assisted scientific computing.Beyond that, get ready to discover the person behind the science – her inspirations, aspirations, and maybe even what she does when not decoding the complexities of machine learning algorithms!Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !Thank you to my Patrons for making this episode possible!Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, Tim Gasser, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Steven Rowland, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady, Kurt TeKolste, Gergely Juhasz, Marcus Nölke, Maggi Mackintosh, Grant Pezzolesi, Avram Aelony, Joshua Meehl, Javier Sabio, Kristian Higgins, Alex Jones, Gregorio Aguilar, Matt Rosinski, Bart Trudeau, Luis Fonseca, Dante Gates, Matt Niccolls, Maksim Kuznecov, Michael Thomas, Luke Gorrie and Cory Kiser.Visit https://www.patreon.com/learnbayesstats to unlock exclusive...
We talked about: Rob's background Going from software engineering to Bayesian modeling Frequentist vs Bayesian modeling approach About integrals Probabilistic programming and samplers MCMC and Hakaru Language vs library Encoding dependencies and relationships into a model Stan, HMC (Hamiltonian Monte Carlo) , and NUTS Sources for learning about Bayesian modeling Reaching out to Rob Links: Book 1: https://bayesiancomputationbook.com/welcome.html Book/Course: https://xcelab.net/rm/statistical-rethinking/ Free ML Engineering course: http://mlzoomcamp.com Join DataTalks.Club: https://datatalks.club/slack.html Our events: https://datatalks.club/events.html
Sarah Minson, U.S. Geological Survey There are many underdetermined geophysical inverse problems. For example, when we try to infer earthquake fault slip, we find that there are many potential slip models that are consistent with our observations and our understanding of earthquake physics. One way to approach these problems is to use Bayesian analysis to infer the ensemble of all potential models that satisfy the observations and our prior knowledge. In Bayesian analysis, our prior knowledge is known as the prior probability density function or prior PDF, the fit to the data is the data likelihood function, and the target PDF that satisfies both the prior PDF and data likelihood function is the posterior PDF. Simulating a posterior PDF can be computationally expensive. Typical earthquake rupture models with 10 km spatial resolution can require using Markov Chain Monte Carlo (MCMC) to draw tens of billions of random realizations of fault slip. And now new technological advancements like LiDAR provide enormous numbers of laser point returns that image surface deformation at submeter scale, exponentially increasing computational cost. How can we make MCMC sampling efficient enough to simulate fault slip distributions at sub-meter scale using “Big Data”? We present a new MCMC approach called cross-fading in which we transition from an analytical posterior PDF (obtained from a conjugate prior to the data likelihood function) to the desired target posterior PDF by bringing in our physical constraints and removing the conjugate prior. This approach has two key efficiencies. First, the starting PDF is by construction “close” to the target posterior PDF, requiring very little MCMC to update the samples to match the target. Second, all PDFs are defined in model space, not data space. The forward model and data misfit are never evaluated during sampling, allowing models to be fit to Big Data with zero computational cost. It is even possible, without additional computational cost, to incorporate model prediction errors for Big Data, that is, to quantify the effects on data prediction of uncertainties in the model design. While we present earthquake models, this approach is flexible and can be applied to many geophysical problems.
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: A case for AI alignment being difficult, published by Jessica Taylor on December 31, 2023 on The AI Alignment Forum. This is an attempt to distill a model of AGI alignment that I have gained primarily from thinkers such as Eliezer Yudkowsky (and to a lesser extent Paul Christiano), but explained in my own terms rather than attempting to hew close to these thinkers. I think I would be pretty good at passing an ideological Turing test for Eliezer Yudowsky on AGI alignment difficulty (but not AGI timelines), though what I'm doing in this post is not that, it's more like finding a branch in the possibility space as I see it that is close enough to Yudowsky's model that it's possible to talk in the same language. Even if the problem turns out to not be very difficult, it's helpful to have a model of why one might think it is difficult, so as to identify weaknesses in the case so as to find AI designs that avoid the main difficulties. Progress on problems can be made by a combination of finding possible paths and finding impossibility results or difficulty arguments. Most of what I say should not be taken as a statement on AGI timelines. Some problems that make alignment difficult, such as ontology identification, also make creating capable AGI difficult to some extent. Defining human values If we don't have a preliminary definition of human values, it's incoherent to talk about alignment. If humans "don't really have values" then we don't really value alignment, so we can't be seriously trying to align AI with human values. There would have to be some conceptual refactor of what problem even makes sense to formulate and try to solve. To the extent that human values don't care about the long term, it's just not important (according to the values of current humans) how the long-term future goes, so the most relevant human values are the longer-term ones. There are idealized forms of expected utility maximization by brute-force search. There are approximations of utility maximization such as reinforcement learning through Bellman equations, MCMC search, and so on. I'm just going to make the assumption that the human brain can be well-modeled as containing one or more approximate expected utility maximizers. It's useful to focus on specific branches of possibility space to flesh out the model, even if the assumption is in some ways problematic. Psychology and neuroscience will, of course, eventually provide more details about what maximizer-like structures in the human brain are actually doing. Given this assumption, the human utility function(s) either do or don't significantly depend on human evolutionary history. I'm just going to assume they do for now. I realize there is some disagreement about how important evopsych is for describing human values versus the attractors of universal learning machines, but I'm going to go with the evopsych branch for now. Given that human brains are well-modeled as containing one or more utility functions, either they're well-modeled as containing one (perhaps which is some sort of monotonic function of multiple other score functions), or it's better to model them as multiple. See shard theory. The difference doesn't matter for now, I'll keep both possibilities open. Eliezer proposes "boredom" as an example of a human value (which could either be its own shard or a term in the utility function). I don't think this is a good example. It's fairly high level and is instrumental to other values. I think "pain avoidance" is a better example due to the possibility of pain asymbolia. Probably, there is some redundancy in the different values (as there is redundancy in trained neural networks, so they still perform well when some neurons are lesioned), which is part of why I don't agree with the fragility of value thesis as stated by Yudkowsky. Re...
It's In the News, a look at the top stories and headlines from the diabetes community happening now. Top stories this week: a new study looks at food-as-medicine for type 2, another FDA warning about fake Ozempic, new research says gut markers may help predict who Tzield will work best for, JDRF partners with NFL and more... Happy New Year - we'll see you in 2024! Find out more about Moms' Night Out Please visit our Sponsors & Partners - they help make the show possible! Take Control with Afrezza Omnipod - Simplify Life Learn about Dexcom Edgepark Medical Supplies Check out VIVI Cap to protect your insulin from extreme temperatures Learn more about AG1 from Athletic Greens Drive research that matters through the T1D Exchange The best way to keep up with Stacey and the show is by signing up for our weekly newsletter: Sign up for our newsletter here Here's where to find us: Facebook (Group) Facebook (Page) Instagram Twitter Check out Stacey's books! Learn more about everything at our home page www.diabetes-connections.com Reach out with questions or comments: info@diabetes-connections.com Episode transcription: Hello and welcome to Diabetes Connections In the News! I'm Stacey Simms and every other Friday I bring you a short episode with the top diabetes stories and headlines happening now. XX In the news is brought to you by Edgepark simplify your diabetes journey with Edgepark XX Our top story this week… XX You often hear people say food is medicine.. but an intensive program trying to show that's the case did NOT improve glycemic control in adults with type 2 diabetes any better than usual care. This was a randomized clinical trial. After 6 months, both groups had a similar drop in HbA1c -- 1.5 percentage points among program enrollees and 1.3 percentage points with usual care, with no significant differences in other metabolic lab values between the groups either, the researchers wrote in JAMA the food-as-medicine participants even gained some weight compared with the usual care group over 6 months (adjusted mean difference 1.95 kg, P=0.04). "I was surprised by the findings because the program is so intensive," Doyle told MedPage Today. "The health system built brick-and-mortar clinics, staffed them with a dietitian, nurse, and community health worker, had weekly food pick-up for 10 meals per week for the entire family, and participants spend a year in the program." Costing an estimated $2,000 annually per participant, the food-as-medicine program allowed participants to choose from a variety of vegetables, fruits, and entrees each week -- enough food for two meals a day, 5 days a week. They were also provided recipes and cooking instructions and met with dietitians to track goals. On the other hand, the control group was only provided usual care, a list of local food bank locations, and the option to join the program after 6 months. The trial was conducted at two sites, one rural and one urban, in the mid-Atlantic region. It recruited 465 adults with type 2 diabetes who completed the study, all of whom started with an HbA1c of 8% or higher. All participants were also self-reported as food insecure. The average age was 54.6 years, 54.8% of participants were female, 81.3% were white, and most resided in the urban location. Of note, all participants also resided in the program's service area and were affiliated with the health system that ran it. "One study should not be over-interpreted," said Doyle. "It is possible that such a program could work in other contexts, among patients less connected to a health system, or in other formats. The main alternative to providing healthy groceries and education is to provide pre-made 'medically tailored meals.'" "I hope the study raises awareness of the potential for food-as-medicine programs to increase healthcare engagement and to push researchers and policymakers to generate more evidence on ways such programs can improve health." It's worth noting that there is very little study – much less clinical trial level study on this type of thing. The researchers say they hope it spurs more research to find methods that will have a large impact. https://news.mit.edu/2023/food-medicine-diabetes-study-1227 https://www.medpagetoday.com/primarycare/dietnutrition/107998 XX New information about moderate low carb diets for people with type 1. The study published in The Lancet Regional Health - Europe is the largest of its kind to date. Participants were for different periods randomly assigned in a crossover manner to eat a traditional diet with 50% of the energy from carbohydrates, or a moderate low-carbohydrate diet with 30% of the energy from carbohydrates. The 50 participants all had type 1 diabetes with elevated mean glucose, long-term blood sugar, and injection therapy with insulin or an insulin pump. Half were women, half men. The average age was 48 years. Participants on a moderate low-carbohydrate diet were found to spend more time in what is known as the target range, the range within which people with type 1 diabetes should be in terms of glucose levels. The increase in time within the target range was an average of 68 minutes per day compared to the traditional diet, while the time with elevated values was reduced by 85 minutes per day. The researchers saw no evidence of adverse effects. https://www.news-medical.net/news/20231220/Moderate-low-carb-diet-safe-and-effective-for-adults-with-type-1-diabetes.aspx XX Researchers at Case Western Reserve University and University Hospitals have identified an enzyme that blocks insulin produced in the body—a discovery that could provide a new target to treat diabetes. Their study, published Dec. 5 in the journal Cell, focuses on nitric oxide, a compound that dilates blood vessels, improves memory, fights infection and stimulates the release of hormones, among other functions. How nitric oxide performs these activities had long been a mystery. The researchers discovered a novel “carrier” enzyme (called SNO-CoA-assisted nitrosylase, or SCAN) that attaches nitric oxide to proteins, including the receptor for insulin action. Given the discovery, next steps could be to develop medications against the enzyme, he said. https://thedaily.case.edu/new-cause-of-diabetes-discovered-offering-potential-target-for-new-classes-of-drugs-to-treat-the-disease/ XX The Food and Drug Administration on Thursday warned consumers not to use counterfeit versions of Novo Nordisk's diabetes drug Ozempic that have been found in the country's drug supply chain. The FDA said it will continue to investigate counterfeit Ozempic 1 milligram injections and has seized thousands of units, but flagged that some may still be available for purchase. The agency said the needles from the seized injections are counterfeit and their sterility cannot be confirmed, which presents an additional risk of infection for patients. Other confirmed counterfeit components from the seized products include the pen label and accompanying information about the healthcare professional and patient, as well as the carton. The FDA urged drug distributors, retail pharmacies, healthcare practitioners and patients to check the drug they have received and to not distribute, use or sell the units labeled with lot number NAR0074 and serial number 430834149057. People who have Ozempic injections with the above lot number and serial number can report it directly to the FDA Office of Criminal Investigations. https://www.nbcnews.com/health/health-news/fda-warns-ozempic-counterfeit-diabetes-weight-loss-rcna130871 XX New research indicates that information in the gut may predict how well a person responds to Tzield. That's the medication approved earlier this year to delay the onset of type 1. These findings reported in the journal Science Translational Medicine, casts a new spotlight on the immune system's relationship with the microbiome, revealing how gut microbes can shape the progression of type 1 diabetes. With this new knowledge in hand, clinicians may better pinpoint patients who are most likely to respond to teplizumab. https://medicalxpress.com/news/2023-12-gut-microbes-patients-response-drug.html XX Experts are advocating for universal screening for type 1 diabetes. With the availability of Tzield and other medications on the horizon, there's a stronger push for screening earlier in life. At least 85% of people who are newly diagnosed do not have a family history of diabetes. Testing for autoantibodies can be completed at home through the TrialNet clinical trial program, or at a doctor's office or lab. For instance, JDRF's T1Detect program provides at-home testing for $55, with lower-cost options for people in financial need. The 2024 American Diabetes Association (ADA) Standards of Care recommend more intensive monitoring for the progression of preclinical type 1 diabetes. The Standards of Care also recommend using Tzield to delay the onset of diabetes in people at least 8 years old with stage 2 type 1 diabetes. https://diatribe.org/type-1-diabetes-it%E2%80%99s-time-population-wide-screening XX Commercial XX https://www.healthline.com/health-news/the-years-biggest-medical-advancements-in-diabetes-treatment XX DRF, the leading global funder of type 1 diabetes (T1D) research, is recognizing the NFL stars who showcased their creativity and a remarkable show of support as part of the highly anticipated annual "My Cause My Cleats" (MCMC) campaign. The My Cause My Cleats initiative allows NFL players to wear custom-painted cleats during selected games to raise awareness and funds for the charitable causes closest to their hearts. The unofficial start of the campaign begins on Giving Tuesday with unboxing day events showcasing the players' cleats and the stories behind them. It continues through weeks 13 and 14 of the season, culminating with the players donning their cleats on game day. After the games, some players donate their cleats to their chosen charities or the NFL auction, with all proceeds going toward their selected causes. Type 1 Diabetes is a life-threatening autoimmune condition that affects people of all ages, regardless of family history or lifestyle choices. To live, people with T1D must carefully balance injecting or infusing insulin with their carbohydrate intake throughout the day and night. T1D impacts approximately 1.6 million people in the U.S. It is unpreventable, and there is currently no cure. This year, JDRF is thankful for the support of several players who have T1D or are advocating for their loved ones with T1D, including Mark Andrews of the Baltimore Ravens, Orlando Brown, Jr. of the Cincinnati Bengals, Blake Ferguson of the Miami Dolphins, Collin Johnson of the Chicago Bears, Chad Muma of the Jacksonville Jaguars, Nate Peterman of the Chicago Bears, and Kevin Radar of the Tennessee Titans. "The NFL players who support JDRF through the My Cause My Cleats exemplify the passion and determination at the heart of the type 1 diabetes community," said Kenya Felton, JDRF Director of PR and Celebrity Engagement. "They serve as inspirations for many adults and children affected by T1D, demonstrating that with an understanding of T1D, effective management, and a good support system, you can overcome the challenges of the disease. Their support helps to increase awareness and is significant in helping JDRF advance life-changing breakthroughs in T1D research and advocacy initiatives." Since its inception in 2016, the MCMC campaign has provided a platform for many NFL players and affiliates to support JDRF's mission, including Beau Benzschawel, David Carr, Will Clarke, Keion Crossen, DeAndre Carter, Reid Ferguson, Jaedan Graham, Jarvis Jenkins, Collin Johnson, Henry Mondeaux, Jaelan Phillips, Adam Schefter, Brandon Wilds, and Jonah Williams. https://www.prnewswire.com/news-releases/nfl-stars-support-jdrf-and-champion-type-1-diabetes-awareness-through-the-my-cause-my-cleats-campaign-302022060.html XX Join us again soon!
The Malaysian Communications and Multimedia Commission (MCMC) is expected to come up with a framework soon to facilitate the registration and regulation of social media platform providers. We discuss how this framework could affect the industry with Alexander Wong, managing editor and co-founder of SoyaCincau.Image credit: MCMC
【12月16日はスペシャルデー】★どちらも無料、参加申し込みは不要です★11時から東京・下北沢でPodcast weekendに出展https://podcastweekend.jp/ 19時からはオンラインで忘年会!https://www.youtube.com/watch?v=pAHguuT21m0 【番組内容】今回の議題は……◆ポキ-1グランプリと同じ井岡と橋本がMCです(0:00)◆朝リスさんのるつぼ相談員は(08:50)◆井岡と橋本の悩みのるつぼ(12:50)◆ポキー1グランプリのプレゼンターは(21:40)◆Podcast weekendのグッズは盛りだくさん(28:54)メルマガ登録はこちらから→ https://bit.ly/asapoki_newsletter ◆朝リスさんからのお便り紹介(35:15)◆東京メンバーとのコール&レスポンス(44:25) ※2023年11月24日に収録しました。広告が入っているときは説明文の時間表記がズレる場合があります。これまでの制作会議はこちら( https://buff.ly/3XER8Co )。 【関連記事】自然に還るプラ、需要急増の見通し 車部品やWiFiルーターも試作https://www.asahi.com/articles/ASRCK4HRWRB0DIFI001.html?iref=omny 【出演・スタッフ】岸上渉橋本佳奈(MC)井岡諒(MC・音源編集) 【朝ポキ情報】ご感想はおたよりフォーム → https://bit.ly/asapoki_otayori 番組カレンダー→ https://bit.ly/asapki_calendar 出演者名検索ツール→ https://bit.ly/asapoki_cast 最新情報はX(旧ツイッター)→ https://bit.ly/asapoki_twitter 交流はコミュニティ → https://bit.ly/asapoki_community テロップ付きはYouTube → https://bit.ly/asapoki_youtube_ こぼれ話はメルマガ → https://bit.ly/asapoki_newsletter 全話あります公式サイト → https://bit.ly/asapoki_lp メールはこちら → podcast@asahi.com See omnystudio.com/listener for privacy information.
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Learning coefficient estimation: the details, published by Zach Furman on November 16, 2023 on LessWrong. What this is for The learning coefficient (LC), or RLCT, is a quantity from singular learning theory that can help to quantify the "complexity" of deep learning models, among other things. This guide is primarily intended to help people interested in improving learning coefficient estimation get up to speed with how it works, behind the scenes. If you're just trying to use the LC for your own project, you can just use the library without knowing all the details, though this guide might still be helpful. It's highly recommended you read this post before reading this one, if you haven't already. We're primarily covering the WBIC paper (Watanabe 2010), the foundation for current LC estimation techniques, but the presentation here is original, aiming for better intuition, and differs substantially from the paper. We'll also briefly cover Lau et al. 2023. Despite all the lengthy talk, what you end up doing in practice is really simple, and the code is designed to highlight that. After some relatively quick setup, the actual LC calculation can be comfortably done in one or two lines of code. What this isn't for A good overview of SLT, or motivation behind studying the LC or loss landscape volume in the first place. We're narrowly focused on LC estimation here. Sampling details. These are very important! But they're not really unique to singular learning theory, and there are plenty of good resources and tutorials on MCMC elsewhere. Derivations of formulas, beyond the high-level reasoning. TLDR What is the learning coefficient? (Review from last time) The learning coefficient (LC), also called the RLCT, measures basin broadness. This isn't new, but typically "basin broadness" is operationalized as "basin flatness" - that is, via the determinant of the Hessian. When the model is singular (eigenvalues of the Hessian are zero), this is a bad idea. The LC operationalizes "basin broadness" as the (low-loss asymptotic) volume scaling exponent. This ends up being the right thing to measure, as justified by singular learning theory. How do we measure it? It turns out that measuring high-dimensional volume directly is hard. We don't do this. Instead we use MCMC to do what's known in statistics as "method of moments" estimation. We contrive a distribution with the LC as a population parameter, sample from that distribution and calculate one of its moments, and solve for the LC. We simplify some details in this section, but this is the conceptual heart of LC estimation. How do we measure it (for real)? The above is a bit simplified. The LC does measure loss volume scaling, but the "loss" it uses is the average or "infinite-data" limit of the empirical loss function. In practice, you don't know this infinite-data loss function. Luckily, you already have a good estimate of it - your empirical loss function. Unluckily, this estimate isn't perfect - it can have some noise. And it turns out this noise is actually worst in the place you least want it. But it all works out in the end! You actually just need to make one small modification to the "idealized" algorithm, and things work fine. This gets you an algorithm that really works in practice! Finally, the state-of-the-art method (Lau et al. 2023) makes a couple simple modifications, for scalability among other reasons: it measures the learning coefficient only *locally*, and uses mini-batch loss instead of full-batch. In chart form: as we move from idealized (top) to realistic (bottom), we get new problems, solutions, and directions for improvement. The guide itself covers the first two rows in the most detail, which are likely the most conceptually difficult to think about, and skips directly from the second row to the fourth row at ...
Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!My Intuitive Bayes Online Courses1:1 Mentorship with meWhat's the difference between MCMC and Variational Inference (VI)? Why is MCMC called an approximate method? When should we use VI instead of MCMC?These are some of the captivating (and practical) questions we'll tackle in this episode. I had the chance to interview Charles Margossian, a research fellow in computational mathematics at the Flatiron Institute, and a core developer of the Stan software.Charles was born and raised in Paris, and then moved to the US to pursue a bachelor's degree in physics at Yale university. After graduating, he worked for two years in biotech, and went on to do a PhD in statistics at Columbia University with someone named… Andrew Gelman — you may have heard of him.Charles is also specialized in pharmacometrics and epidemiology, so we also talked about some practical applications of Bayesian methods and algorithms in these fascinating fields.Oh, and Charles' life doesn't only revolve around computers: he practices ballroom dancing and pickup soccer, and used to do improvised musical comedy!Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !Thank you to my Patrons for making this episode possible!Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, Tim Gasser, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Raul Maldonado, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Steven Rowland, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Trey Causey, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady, Kurt TeKolste, Gergely Juhasz, Marcus Nölke, Maggi Mackintosh, Grant Pezzolesi, Avram Aelony, Joshua Meehl, Javier Sabio, Kristian Higgins, Alex Jones, Gregorio Aguilar and Matt Rosinski.Visit https://www.patreon.com/learnbayesstats to unlock exclusive Bayesian swag ;)Links from the show:Charles' website: https://charlesm93.github.io/Charles on Twitter: https://twitter.com/charlesm993Charles on GitHub:
7月はAny Given Saturdayのポッドキャスト強化月間です!強化月間最後のエピソードはゲスト回!!今回はNFLやカレッジフットボールの現地観戦する達人、MC79さん(@_MC79)さんにお越しいただきました! ✅MCさんのフットボール遍歴 ✅MCさんが選ぶカレッジスタジアム5選 ✅現地観戦のコツ(チケット購入方法など) 是非ご試聴ください!! ========================= ◆エピソード#000【トレーラー】 トレーラーのくせに30分と長いのですが(苦笑)初めての方 は是非本編の前に聴いてみてください! ◆ウェイブサイト「Any Given Saturday」 アメリカにおいてアマチュアスポーツながら絶大なる人気を誇るカレッジフットボール(大学アメフト)の様々な情報を紹介するサイト。カレッジフットボールを楽しむための基礎知識からシーズン中の試合分析、スコア、ランキングなどの紹介。加えて大学選手がプロ入りするために避けて通れないNFLドラフトの情報も。 ◆インスタグラム ◆フェイスブック ◆スタンドFM 主にライブ配信形式でカレッジフットボールの事をフリースタイルで話しています。そのアーカイブ音源を試聴していただけます。 ◆匿名で質問したい方はこちらのGoogle Formsでどうぞ ◆素材提供 MaxKoMusic | https://maxkomusic.com/ Chosic | https://www.chosic.com/free-music/all/ MusMus | https://musmus.main/jp/ Pixabay | https://pixabay.com/music/ #カレッジフットボール #アメリカンフットボール #アメフト
1. AG Systems - Active Tekno 2. Helix & DJ Fury - Insane Asylum 3. AG Systems - Unda Ground 4. Square Wave - Like A Criminal 5. Helix & Tekno Dred - Mindless Pleasure 6. Helix - Now Control 7. Tekno Dred & Ad Man - A Voice Spoke To Me (Helix Remix) 8. MC MC & Rush Hour - Music Maker (Fury's Vengeance Mix) 9. Square Wave - God's Broth 10. DJ Fury - Lemonade Raygun (Back Room Mix) 11. Sharkey & Marc Smith - Utopia 12. Marc Smith - Nothing More 13. Marc Smith - Procrastinator
Communications and Digital Minister Fahmi Fadzil recently called out social media platform Telegram for not taking action on illicit activity on the platform. Does the MCMC have the authority and legal means to regulate digital platforms hosted overseas? We discuss the legal framework that applies and gaps in regulation with Intellectual Property and Information Technology Lawyer, Foong Cheng Leong.Image by: Shutterstock
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: The Lightcone Theorem: A Better Foundation For Natural Abstraction?, published by johnswentworth on May 15, 2023 on The AI Alignment Forum. Credit to David Lorell for serving as an active sounding board as the ideas in this post were developed. For about a year and a half now, my main foundation for natural abstraction math has been The Telephone Theorem: long-range interactions in a probabilistic graphical model (in the long-range limit) are mediated by quantities which are conserved (in the long-range limit). From there, the next big conceptual step is to argue that the quantities conserved in the long-range limit are also conserved by resampling, and therefore the conserved quantities of an MCMC sampling process on the model mediate all long-range interactions in the model. The most immediate shortcoming of the Telephone Theorem and the resampling argument is that they talk about behavior in infinite limits. To use them, either we need to have an infinitely large graphical model, or we need to take an approximation. For practical purposes, approximation is clearly the way to go, but just directly adding epsilons and deltas to the arguments gives relatively weak results. This post presents a different path. The core result is the Lightcone Theorem: Start with a probabilistic graphical model on the variables X1,.,Xn. The graph defines adjacency, distance, etc between variables. For directed graphical models (i.e. Bayes nets), spouses (as well as parents and children) count as adjacent. We can model those variables as the output of a Gibbs sampler (that's the MCMC process) on the graphical model. Call the initial condition of the sampler X0=(X01,.,X0n). The distribution of X0 must be the same as the distribution of X (i.e. the sampler is initialized “in equilibrium”). We can model the sampler as having run for any number of steps to generate the variables; call the number of steps T. At each step, the process resamples some set of nonadjacent variables conditional on their neighbors. The Lightcone Theorem says: conditional on X0, any sets of variables in X which are a distance of at least 2T apart in the graphical model are independent. Yes, exactly independent, no approximation. In short: the initial condition of the resampling process provides a latent, conditional on which we have exact independence at a distance. This was. rather surprising to me. If you'd floated the Lightcone Theorem as a conjecture a year ago, I'd have said it would probably work as an approximation for large T, but no way it would work exactly for finite T. Yet here we are. The Proof, In Pictures The proof is best presented visually. High-level outline: Perform a do() operation on the Gibbs sampler, so that it never resamples the variables a distance of T from XR. In the do()-operated process, X0 mediates between XTR and XTD(R,≥2T), where D(R,≥2T) indicates indices of variables a distance of at least 2T from XR. Since X0, XTR and XTD(R,≥2T) are all outside the lightcone of the do()-operation, they have the same joint distribution under the non-do()-operated sampler as under the do()-operated sampler. Therefore X0 mediates between XTR and XTD(R,≥2T) under the original sampler. We start with the graphical model: Within that graphical model, we'll pick some tuple of variables XR (“R” for “region”). I'll use the notation XD(R,t) for the variables a distance t away from R, XD(R,>t) for variables a distance greater than t away from R, XD(R,
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: The Lightcone Theorem: A Better Foundation For Natural Abstraction?, published by johnswentworth on May 15, 2023 on LessWrong. Credit to David Lorell for serving as an active sounding board as the ideas in this post were developed. For about a year and a half now, my main foundation for natural abstraction math has been The Telephone Theorem: long-range interactions in a probabilistic graphical model (in the long-range limit) are mediated by quantities which are conserved (in the long-range limit). From there, the next big conceptual step is to argue that the quantities conserved in the long-range limit are also conserved by resampling, and therefore the conserved quantities of an MCMC sampling process on the model mediate all long-range interactions in the model. The most immediate shortcoming of the Telephone Theorem and the resampling argument is that they talk about behavior in infinite limits. To use them, either we need to have an infinitely large graphical model, or we need to take an approximation. For practical purposes, approximation is clearly the way to go, but just directly adding epsilons and deltas to the arguments gives relatively weak results. This post presents a different path. The core result is the Lightcone Theorem: Start with a probabilistic graphical model on the variables X1,.,Xn. The graph defines adjacency, distance, etc between variables. For directed graphical models (i.e. Bayes nets), spouses (as well as parents and children) count as adjacent. We can model those variables as the output of a Gibbs sampler (that's the MCMC process) on the graphical model. Call the initial condition of the sampler X0=(X01,.,X0n). The distribution of X0 must be the same as the distribution of X (i.e. the sampler is initialized “in equilibrium”). We can model the sampler as having run for any number of steps to generate the variables; call the number of steps T. At each step, the process resamples some set of nonadjacent variables conditional on their neighbors. The Lightcone Theorem says: conditional on X0, any sets of variables in X which are a distance of at least 2T apart in the graphical model are independent. Yes, exactly independent, no approximation. In short: the initial condition of the resampling process provides a latent, conditional on which we have exact independence at a distance. This was. rather surprising to me. If you'd floated the Lightcone Theorem as a conjecture a year ago, I'd have said it would probably work as an approximation for large T, but no way it would work exactly for finite T. Yet here we are. The Proof, In Pictures The proof is best presented visually. High-level outline: Perform a do() operation on the Gibbs sampler, so that it never resamples the variables a distance of T from XR. In the do()-operated process, X0 mediates between XTR and XTD(R,≥2T), where D(R,≥2T) indicates indices of variables a distance of at least 2T from XR. Since X0, XTR and XTD(R,≥2T) are all outside the lightcone of the do()-operation, they have the same joint distribution under the non-do()-operated sampler as under the do()-operated sampler. Therefore X0 mediates between XTR and XTD(R,≥2T) under the original sampler. We start with the graphical model: Within that graphical model, we'll pick some tuple of variables XR (“R” for “region”). I'll use the notation XD(R,t) for the variables a distance t away from R, XD(R,>t) for variables a distance greater than t away from R, XD(R,
Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!My Intuitive Bayes Online Courses1:1 Mentorship with me------------------------------------------------------------------------------Max Kochurov's State of Bayes Lecture Series: https://www.youtube.com/playlist?list=PL1iMFW7frOOsh5KOcfvKWM12bjh8zs9BQSign up here for upcoming lessons: https://www.meetup.com/pymc-labs-online-meetup/events/293101751/------------------------------------------------------------------------------We talk a lot about different MCMC methods on this podcast, because they are the workhorses of the Bayesian models. But other methods exist to infer the posterior distributions of your models — like Sequential Monte Carlo (SMC) for instance. You've never heard of SMC? Well perfect, because Nicolas Chopin is gonna tell you all about it in this episode!A lecturer at the French university of ENSAE since 2006, Nicolas is one of the world experts on SMC. Before that, he graduated from Ecole Polytechnique and… ENSAE, where he did his PhD from 1999 to 2003.Outside of work, Nicolas enjoys spending time with his family, practicing aikido, and reading a lot of books.Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !Thank you to my Patrons for making this episode possible!Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, Tim Gasser, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Nathaniel Neitzke, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Raul Maldonado, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Steven Rowland, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Trey Causey, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady and Kurt TeKolste.Visit https://www.patreon.com/learnbayesstats to unlock exclusive Bayesian swag ;)Links from the show:Old episodes...
Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!Matt Hoffman has already worked on many topics in his life – music information retrieval, speech enhancement, user behavior modeling, social network analysis, astronomy, you name it.Obviously, picking questions for him was hard, so we ended up talking more or less freely — which is one of my favorite types of episodes, to be honest.You'll hear about the circumstances Matt would advise picking up Bayesian stats, generalized HMC, blocked samplers, why do the samplers he works on have food-based names, etc.In case you don't know him, Matt is a research scientist at Google. Before that, he did a postdoc in the Columbia Stats department, working with Andrew Gelman, and a Ph.D at Princeton, working with David Blei and Perry Cook.Matt is probably best known for his work in approximate Bayesian inference algorithms, such as stochastic variational inference and the no-U-turn sampler, but he's also worked on a wide range of applications, and contributed to software such as Stan and TensorFlow Probability.Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !Thank you to my Patrons for making this episode possible!Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, Tim Gasser, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor, Thomas Wiecki, Chad Scherrer, Nathaniel Neitzke, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Joshua Duncan, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Raul Maldonado, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Steven Rowland, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, David Haas, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Trey Causey, Andreas Kröpelin, Raphaël R, Nicolas Rode and Gabriel Stechschulte.Visit https://www.patreon.com/learnbayesstats to unlock exclusive Bayesian swag ;)Links from the show:Matt's website: http://matthewdhoffman.com/Matt on Google Scholar: https://scholar.google.com/citations?hl=en&user=IeHKeGYAAAAJ&view_op=list_worksThe No-U-Turn Sampler: Adaptively Setting Path Lengths in Hamiltonian Monte Carlo: https://www.jmlr.org/papers/volume15/hoffman14a/hoffman14a.pdfTuning-Free Generalized Hamiltonian Monte Carlo:
En este episodio de Entre Amigos, Rebeca, Victor, y Carlos nos ayudan con el análisis del juego entre los Chiefs y Broncos de la Semana 14. Todas las ganancias de Damani's MCMC se destinarán a la Liga de Futbol Americano en Monterrey. Haga su oferta aquí https://rb.gy/nc0902 Les recordamos votar por nuestro Fan of The Year. Vota por Juan aquí https://bit.ly/3g3trCiSee omnystudio.com/listener for privacy information.
Today we sat down with SSgt Andrew Dow about his experience in MCMC NCR and how it catapulted him into his passion for practical shooting since his first MCMC.
你的周舟,她现在在哪里。 时间轴: 14:25 MC黄上大学时的嫂子带给MC黄的心灵震撼 55:43 台长把自己的照片传到了“北京月老赐婚”上 音乐: 1. 狼狈 - 康姆士COM‘Z 本期主播:小鲍、小黄 喜欢我们快快搜:radio578,加入【578广播花好月圆俱乐部】,收获更多快乐
你的周舟,她现在在哪里。 时间轴: 14:25 MC黄上大学时的嫂子带给MC黄的心灵震撼 55:43 台长把自己的照片传到了“北京月老赐婚”上 音乐: 1. 狼狈 - 康姆士COM‘Z 本期主播:小鲍、小黄 喜欢我们快快搜:radio578,加入【578广播花好月圆俱乐部】,收获更多快乐
-Introductions -How long have you been a competitive shooter? -What disciplines do you shoot? -Can you talk about how you got into the tactical games? -How did you and Frank prepare for the WV Games in terms of shooting and fitness? -Recap of the events: Axel Bar two-man carry Farmer's carry while partner is shooting, and sandbag over yoke 5-mile run Yoke, Sandbag, Farmer Carry Sled, sandbag, husafell stone relay El Cartel -How do you feel you did, based on your level of preparation and experience? -How do you recommend Marines who are interested in Tactical Games prepare (timeline, shooting, body maintenance, game day maintenance) -You shot your first MCMC at Stone Bay, any thoughts on those two weeks? -What are some lessons you learned from TTG WV and MCMC East that you are going to incorporate into your personal fitness and shooting training? -You are PCS'ing to 3rdMarDiv this summer- what are some aspects from TTG and the MCMC that you plan on bringing to your unit? -Anything else you'd like to leave listeners with?
In this episode, I sat down with various Marines who competed in the Marine Corps Marksmanship Competition National Capital Region and got their take on the experience from this year's match. Instagram: @usmcshootingteam @mfgundlach