Podcasts about Generalization

  • 421PODCASTS
  • 568EPISODES
  • 43mAVG DURATION
  • 1WEEKLY EPISODE
  • Feb 13, 2026LATEST
Generalization

POPULARITY

20192020202120222023202420252026


Best podcasts about Generalization

Latest podcast episodes about Generalization

Neuro Navigators: A MedBridge Podcast
Neuro Navigators Episode 24: Can Motor Performance Be Driven By Cognition? The CO-OP Approach

Neuro Navigators: A MedBridge Podcast

Play Episode Listen Later Feb 13, 2026 56:27


Helene Polatajko, PhD, OT(C), FCAOT, FCAHS, LLD(h.c.), OC, a renowned occupational therapist, researcher, and co-developer of the CO-OP Approach, joins host J.J. Mowder-Tinney for a thought-provoking conversation on how cognition can drive motor performance. Together, they explore the power of guided discovery, client-centered goals, and the surprising role of self-generated strategies in rehabilitation. From dancing to dressing to stroke recovery, you'll hear how thinking differently about movement can change what your patients are capable of. Tune in to reframe your clinical lens and get inspired to incorporate “goal-plan-do-check” into your own sessions.Learning ObjectivesAnalyze the evidence around CO-OP approaches to meaningful activities across pediatric and adult populationsApply evidence-based, practical strategies to actionably address challenges in occupationalSolve patient case scenarios involving clients with coordination or motor learning impairmentTimestamps(00:00:00) Welcome(00:00:05) Introduction to cognition and motor-based performance(00:01:20) Dr. Polatajko's background and journey in occupational therapy(00:05:30) The self-driving car(00:12:40) Cognitive orientation to daily occupational performance (CO-OP)(00:14:35) Dynamic performance analysis(00:22:15) Guided discovery(00:30:24) Generalization and transfer of skills(00:34:14) Goal-plan-do-check(00:53:25) Key takeaways and conclusionNeuro Navigators is brought to you by Medbridge. If you'd like to earn continuing education credit for listening to this episode and access bonus takeaway handouts, log in to your Medbridge account and navigate to the course where you'll find accreditation details. If applicable, complete the post-course assessment and survey to be eligible for credit. The takeaway handout on Medbridge gives you the key points mentioned in this episode, along with additional resources you can implement into your practice right away.To hear more episodes of Neuro Naviagators, visit ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠https://www.medbridge.com/neuro-navigators⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠If you'd like to subscribe to Medbridge, visit ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠https://www.medbridge.com/pricing/⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠IG: ⁠⁠⁠⁠⁠⁠⁠https://www.instagram.com/medbridgeteam/

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0
The First Mechanistic Interpretability Frontier Lab — Myra Deng & Mark Bissell of Goodfire AI

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

Play Episode Listen Later Feb 6, 2026 68:01


From Palantir and Two Sigma to building Goodfire into the poster-child for actionable mechanistic interpretability, Mark Bissell (Member of Technical Staff) and Myra Deng (Head of Product) are trying to turn “peeking inside the model” into a repeatable production workflow by shipping APIs, landing real enterprise deployments, and now scaling the bet with a recent $150M Series B funding round at a $1.25B valuation.In this episode, we go far beyond the usual “SAEs are cool” take. We talk about Goodfire's core bet: that the AI lifecycle is still fundamentally broken because the only reliable control we have is data and we post-train, RLHF, and fine-tune by “slurping supervision through a straw,” hoping the model picks up the right behaviors while quietly absorbing the wrong ones. Goodfire's answer is to build a bi-directional interface between humans and models: read what's happening inside, edit it surgically, and eventually use interpretability during training so customization isn't just brute-force guesswork.Mark and Myra walk through what that looks like when you stop treating interpretability like a lab demo and start treating it like infrastructure: lightweight probes that add near-zero latency, token-level safety filters that can run at inference time, and interpretability workflows that survive messy constraints (multilingual inputs, synthetic→real transfer, regulated domains, no access to sensitive data). We also get a live window into what “frontier-scale interp” means operationally (i.e. steering a trillion-parameter model in real time by targeting internal features) plus why the same tooling generalizes cleanly from language models to genomics, medical imaging, and “pixel-space” world models.We discuss:* Myra + Mark's path: Palantir (health systems, forward-deployed engineering) → Goodfire early team; Two Sigma → Head of Product, translating frontier interpretability research into a platform and real-world deployments* What “interpretability” actually means in practice: not just post-hoc poking, but a broader “science of deep learning” approach across the full AI lifecycle (data curation → post-training → internal representations → model design)* Why post-training is the first big wedge: “surgical edits” for unintended behaviors likereward hacking, sycophancy, noise learned during customization plus the dream of targeted unlearning and bias removal without wrecking capabilities* SAEs vs probes in the real world: why SAE feature spaces sometimes underperform classifiers trained on raw activations for downstream detection tasks (hallucination, harmful intent, PII), and what that implies about “clean concept spaces”* Rakuten in production: deploying interpretability-based token-level PII detection at inference time to prevent routing private data to downstream providers plus the gnarly constraints: no training on real customer PII, synthetic→real transfer, English + Japanese, and tokenization quirks* Why interp can be operationally cheaper than LLM-judge guardrails: probes are lightweight, low-latency, and don't require hosting a second large model in the loop* Real-time steering at frontier scale: a demo of steering Kimi K2 (~1T params) live and finding features via SAE pipelines, auto-labeling via LLMs, and toggling a “Gen-Z slang” feature across multiple layers without breaking tool use* Hallucinations as an internal signal: the case that models have latent uncertainty / “user-pleasing” circuitry you can detect and potentially mitigate more directly than black-box methods* Steering vs prompting: the emerging view that activation steering and in-context learning are more closely connected than people think, including work mapping between the two (even for jailbreak-style behaviors)* Interpretability for science: using the same tooling across domains (genomics, medical imaging, materials) to debug spurious correlations and extract new knowledge up to and including early biomarker discovery work with major partners* World models + “pixel-space” interpretability: why vision/video models make concepts easier to see, how that accelerates the feedback loop, and why robotics/world-model partners are especially interesting design partners* The north star: moving from “data in, weights out” to intentional model design where experts can impart goals and constraints directly, not just via reward signals and brute-force post-training—Goodfire AI* Website: https://goodfire.ai* LinkedIn: https://www.linkedin.com/company/goodfire-ai/* X: https://x.com/GoodfireAIMyra Deng* Website: https://myradeng.com/* LinkedIn: https://www.linkedin.com/in/myra-deng/* X: https://x.com/myra_dengMark Bissell* LinkedIn: https://www.linkedin.com/in/mark-bissell/* X: https://x.com/MarkMBissellFull Video EpisodeTimestamps00:00:00 Introduction00:00:05 Introduction to the Latent Space Podcast and Guests from Goodfire00:00:29 What is Goodfire? Mission and Focus on Interpretability00:01:01 Goodfire's Practical Approach to Interpretability00:01:37 Goodfire's Series B Fundraise Announcement00:02:04 Backgrounds of Mark and Myra from Goodfire00:02:51 Team Structure and Roles at Goodfire00:05:13 What is Interpretability? Definitions and Techniques00:05:30 Understanding Errors00:07:29 Post-training vs. Pre-training Interpretability Applications00:08:51 Using Interpretability to Remove Unwanted Behaviors00:10:09 Grokking, Double Descent, and Generalization in Models00:10:15 404 Not Found Explained00:12:06 Subliminal Learning and Hidden Biases in Models00:14:07 How Goodfire Chooses Research Directions and Projects00:15:00 Troubleshooting Errors00:16:04 Limitations of SAEs and Probes in Interpretability00:18:14 Rakuten Case Study: Production Deployment of Interpretability00:20:45 Conclusion00:21:12 Efficiency Benefits of Interpretability Techniques00:21:26 Live Demo: Real-Time Steering in a Trillion Parameter Model00:25:15 How Steering Features are Identified and Labeled00:26:51 Detecting and Mitigating Hallucinations Using Interpretability00:31:20 Equivalence of Activation Steering and Prompting00:34:06 Comparing Steering with Fine-Tuning and LoRA Techniques00:36:04 Model Design and the Future of Intentional AI Development00:38:09 Getting Started in Mechinterp: Resources, Programs, and Open Problems00:40:51 Industry Applications and the Rise of Mechinterp in Practice00:41:39 Interpretability for Code Models and Real-World Usage00:43:07 Making Steering Useful for More Than Stylistic Edits00:46:17 Applying Interpretability to Healthcare and Scientific Discovery00:49:15 Why Interpretability is Crucial in High-Stakes Domains like Healthcare00:52:03 Call for Design Partners Across Domains00:54:18 Interest in World Models and Visual Interpretability00:57:22 Sci-Fi Inspiration: Ted Chiang and Interpretability01:00:14 Interpretability, Safety, and Alignment Perspectives01:04:27 Weak-to-Strong Generalization and Future Alignment Challenges01:05:38 Final Thoughts and Hiring/Collaboration Opportunities at GoodfireTranscriptShawn Wang [00:00:05]: So welcome to the Latent Space pod. We're back in the studio with our special MechInterp co-host, Vibhu. Welcome. Mochi, Mochi's special co-host. And Mochi, the mechanistic interpretability doggo. We have with us Mark and Myra from Goodfire. Welcome. Thanks for having us on. Maybe we can sort of introduce Goodfire and then introduce you guys. How do you introduce Goodfire today?Myra Deng [00:00:29]: Yeah, it's a great question. So Goodfire, we like to say, is an AI research lab that focuses on using interpretability to understand, learn from, and design AI models. And we really believe that interpretability will unlock the new generation, next frontier of safe and powerful AI models. That's our description right now, and I'm excited to dive more into the work we're doing to make that happen.Shawn Wang [00:00:55]: Yeah. And there's always like the official description. Is there an understatement? Is there an unofficial one that sort of resonates more with a different audience?Mark Bissell [00:01:01]: Well, being an AI research lab that's focused on interpretability, there's obviously a lot of people have a lot that they think about when they think of interpretability. And I think we have a pretty broad definition of what that means and the types of places that can be applied. And in particular, applying it in production scenarios, in high stakes industries, and really taking it sort of from the research world into the real world. Which, you know. It's a new field, so that hasn't been done all that much. And we're excited about actually seeing that sort of put into practice.Shawn Wang [00:01:37]: Yeah, I would say it wasn't too long ago that Anthopic was like still putting out like toy models or superposition and that kind of stuff. And I wouldn't have pegged it to be this far along. When you and I talked at NeurIPS, you were talking a little bit about your production use cases and your customers. And then not to bury the lead, today we're also announcing the fundraise, your Series B. $150 million. $150 million at a 1.25B valuation. Congrats, Unicorn.Mark Bissell [00:02:02]: Thank you. Yeah, no, things move fast.Shawn Wang [00:02:04]: We were talking to you in December and already some big updates since then. Let's dive, I guess, into a bit of your backgrounds as well. Mark, you were at Palantir working on health stuff, which is really interesting because the Goodfire has some interesting like health use cases. I don't know how related they are in practice.Mark Bissell [00:02:22]: Yeah, not super related, but I don't know. It was helpful context to know what it's like. Just to work. Just to work with health systems and generally in that domain. Yeah.Shawn Wang [00:02:32]: And Mara, you were at Two Sigma, which actually I was also at Two Sigma back in the day. Wow, nice.Myra Deng [00:02:37]: Did we overlap at all?Shawn Wang [00:02:38]: No, this is when I was briefly a software engineer before I became a sort of developer relations person. And now you're head of product. What are your sort of respective roles, just to introduce people to like what all gets done in Goodfire?Mark Bissell [00:02:51]: Yeah, prior to Goodfire, I was at Palantir for about three years as a forward deployed engineer, now a hot term. Wasn't always that way. And as a technical lead on the health care team and at Goodfire, I'm a member of the technical staff. And honestly, that I think is about as specific as like as as I could describe myself because I've worked on a range of things. And, you know, it's it's a fun time to be at a team that's still reasonably small. I think when I joined one of the first like ten employees, now we're above 40, but still, it looks like there's always a mix of research and engineering and product and all of the above. That needs to get done. And I think everyone across the team is, you know, pretty, pretty switch hitter in the roles they do. So I think you've seen some of the stuff that I worked on related to image models, which was sort of like a research demo. More recently, I've been working on our scientific discovery team with some of our life sciences partners, but then also building out our core platform for more of like flexing some of the kind of MLE and developer skills as well.Shawn Wang [00:03:53]: Very generalist. And you also had like a very like a founding engineer type role.Myra Deng [00:03:58]: Yeah, yeah.Shawn Wang [00:03:59]: So I also started as I still am a member of technical staff, did a wide range of things from the very beginning, including like finding our office space and all of this, which is we both we both visited when you had that open house thing. It was really nice.Myra Deng [00:04:13]: Thank you. Thank you. Yeah. Plug to come visit our office.Shawn Wang [00:04:15]: It looked like it was like 200 people. It has room for 200 people. But you guys are like 10.Myra Deng [00:04:22]: For a while, it was very empty. But yeah, like like Mark, I spend. A lot of my time as as head of product, I think product is a bit of a weird role these days, but a lot of it is thinking about how do we take our frontier research and really apply it to the most important real world problems and how does that then translate into a platform that's repeatable or a product and working across, you know, the engineering and research teams to make that happen and also communicating to the world? Like, what is interpretability? What is it used for? What is it good for? Why is it so important? All of these things are part of my day-to-day as well.Shawn Wang [00:05:01]: I love like what is things because that's a very crisp like starting point for people like coming to a field. They all do a fun thing. Vibhu, why don't you want to try tackling what is interpretability and then they can correct us.Vibhu Sapra [00:05:13]: Okay, great. So I think like one, just to kick off, it's a very interesting role to be head of product, right? Because you guys, at least as a lab, you're more of an applied interp lab, right? Which is pretty different than just normal interp, like a lot of background research. But yeah. You guys actually ship an API to try these things. You have Ember, you have products around it, which not many do. Okay. What is interp? So basically you're trying to have an understanding of what's going on in model, like in the model, in the internal. So different approaches to do that. You can do probing, SAEs, transcoders, all this stuff. But basically you have an, you have a hypothesis. You have something that you want to learn about what's happening in a model internals. And then you're trying to solve that from there. You can do stuff like you can, you know, you can do activation mapping. You can try to do steering. There's a lot of stuff that you can do, but the key question is, you know, from input to output, we want to have a better understanding of what's happening and, you know, how can we, how can we adjust what's happening on the model internals? How'd I do?Mark Bissell [00:06:12]: That was really good. I think that was great. I think it's also a, it's kind of a minefield of a, if you ask 50 people who quote unquote work in interp, like what is interpretability, you'll probably get 50 different answers. And. Yeah. To some extent also like where, where good fire sits in the space. I think that we're an AI research company above all else. And interpretability is a, is a set of methods that we think are really useful and worth kind of specializing in, in order to accomplish the goals we want to accomplish. But I think we also sort of see some of the goals as even more broader as, as almost like the science of deep learning and just taking a not black box approach to kind of any part of the like AI development life cycle, whether that. That means using interp for like data curation while you're training your model or for understanding what happened during post-training or for the, you know, understanding activations and sort of internal representations, what is in there semantically. And then a lot of sort of exciting updates that were, you know, are sort of also part of the, the fundraise around bringing interpretability to training, which I don't think has been done all that much before. A lot of this stuff is sort of post-talk poking at models as opposed to. To actually using this to intentionally design them.Shawn Wang [00:07:29]: Is this post-training or pre-training or is that not a useful.Myra Deng [00:07:33]: Currently focused on post-training, but there's no reason the techniques wouldn't also work in pre-training.Shawn Wang [00:07:38]: Yeah. It seems like it would be more active, applicable post-training because basically I'm thinking like rollouts or like, you know, having different variations of a model that you can tweak with the, with your steering. Yeah.Myra Deng [00:07:50]: And I think in a lot of the news that you've seen in, in, on like Twitter or whatever, you've seen a lot of unintended. Side effects come out of post-training processes, you know, overly sycophantic models or models that exhibit strange reward hacking behavior. I think these are like extreme examples. There's also, you know, very, uh, mundane, more mundane, like enterprise use cases where, you know, they try to customize or post-train a model to do something and it learns some noise or it doesn't appropriately learn the target task. And a big question that we've always had is like, how do you use your understanding of what the model knows and what it's doing to actually guide the learning process?Shawn Wang [00:08:26]: Yeah, I mean, uh, you know, just to anchor this for people, uh, one of the biggest controversies of last year was 4.0 GlazeGate. I've never heard of GlazeGate. I didn't know that was what it was called. The other one, they called it that on the blog post and I was like, well, how did OpenAI call it? Like officially use that term. And I'm like, that's funny, but like, yeah, I guess it's the pitch that if they had worked a good fire, they wouldn't have avoided it. Like, you know what I'm saying?Myra Deng [00:08:51]: I think so. Yeah. Yeah.Mark Bissell [00:08:53]: I think that's certainly one of the use cases. I think. Yeah. Yeah. I think the reason why post-training is a place where this makes a lot of sense is a lot of what we're talking about is surgical edits. You know, you want to be able to have expert feedback, very surgically change how your model is doing, whether that is, you know, removing a certain behavior that it has. So, you know, one of the things that we've been looking at or is, is another like common area where you would want to make a somewhat surgical edit is some of the models that have say political bias. Like you look at Quen or, um, R1 and they have sort of like this CCP bias.Shawn Wang [00:09:27]: Is there a CCP vector?Mark Bissell [00:09:29]: Well, there's, there are certainly internal, yeah. Parts of the representation space where you can sort of see where that lives. Yeah. Um, and you want to kind of, you know, extract that piece out.Shawn Wang [00:09:40]: Well, I always say, you know, whenever you find a vector, a fun exercise is just like, make it very negative to see what the opposite of CCP is.Mark Bissell [00:09:47]: The super America, bald eagles flying everywhere. But yeah. So in general, like lots of post-training tasks where you'd want to be able to, to do that. Whether it's unlearning a certain behavior or, you know, some of the other kind of cases where this comes up is, are you familiar with like the, the grokking behavior? I mean, I know the machine learning term of grokking.Shawn Wang [00:10:09]: Yeah.Mark Bissell [00:10:09]: Sort of this like double descent idea of, of having a model that is able to learn a generalizing, a generalizing solution, as opposed to even if memorization of some task would suffice, you want it to learn the more general way of doing a thing. And so, you know, another. A way that you can think about having surgical access to a model's internals would be learn from this data, but learn in the right way. If there are many possible, you know, ways to, to do that. Can make interp solve the double descent problem?Shawn Wang [00:10:41]: Depends, I guess, on how you. Okay. So I, I, I viewed that double descent as a problem because then you're like, well, if the loss curves level out, then you're done, but maybe you're not done. Right. Right. But like, if you actually can interpret what is a generalizing or what you're doing. What is, what is still changing, even though the loss is not changing, then maybe you, you can actually not view it as a double descent problem. And actually you're just sort of translating the space in which you view loss and like, and then you have a smooth curve. Yeah.Mark Bissell [00:11:11]: I think that's certainly like the domain of, of problems that we're, that we're looking to get.Shawn Wang [00:11:15]: Yeah. To me, like double descent is like the biggest thing to like ML research where like, if you believe in scaling, then you don't need, you need to know where to scale. And. But if you believe in double descent, then you don't, you don't believe in anything where like anything levels off, like.Vibhu Sapra [00:11:30]: I mean, also tendentially there's like, okay, when you talk about the China vector, right. There's the subliminal learning work. It was from the anthropic fellows program where basically you can have hidden biases in a model. And as you distill down or, you know, as you train on distilled data, those biases always show up, even if like you explicitly try to not train on them. So, you know, it's just like another use case of. Okay. If we can interpret what's happening in post-training, you know, can we clear some of this? Can we even determine what's there? Because yeah, it's just like some worrying research that's out there that shows, you know, we really don't know what's going on.Mark Bissell [00:12:06]: That is. Yeah. I think that's the biggest sentiment that we're sort of hoping to tackle. Nobody knows what's going on. Right. Like subliminal learning is just an insane concept when you think about it. Right. Train a model on not even the logits, literally the output text of a bunch of random numbers. And now your model loves owls. And you see behaviors like that, that are just, they defy, they defy intuition. And, and there are mathematical explanations that you can get into, but. I mean.Shawn Wang [00:12:34]: It feels so early days. Objectively, there are a sequence of numbers that are more owl-like than others. There, there should be.Mark Bissell [00:12:40]: According to, according to certain models. Right. It's interesting. I think it only applies to models that were initialized from the same starting Z. Usually, yes.Shawn Wang [00:12:49]: But I mean, I think that's a, that's a cheat code because there's not enough compute. But like if you believe in like platonic representation, like probably it will transfer across different models as well. Oh, you think so?Mark Bissell [00:13:00]: I think of it more as a statistical artifact of models initialized from the same seed sort of. There's something that is like path dependent from that seed that might cause certain overlaps in the latent space and then sort of doing this distillation. Yeah. Like it pushes it towards having certain other tendencies.Vibhu Sapra [00:13:24]: Got it. I think there's like a bunch of these open-ended questions, right? Like you can't train in new stuff during the RL phase, right? RL only reorganizes weights and you can only do stuff that's somewhat there in your base model. You're not learning new stuff. You're just reordering chains and stuff. But okay. My broader question is when you guys work at an interp lab, how do you decide what to work on and what's kind of the thought process? Right. Because we can ramble for hours. Okay. I want to know this. I want to know that. But like, how do you concretely like, you know, what's the workflow? Okay. There's like approaches towards solving a problem, right? I can try prompting. I can look at chain of thought. I can train probes, SAEs. But how do you determine, you know, like, okay, is this going anywhere? Like, do we have set stuff? Just, you know, if you can help me with all that. Yeah.Myra Deng [00:14:07]: It's a really good question. I feel like we've always at the very beginning of the company thought about like, let's go and try to learn what isn't working in machine learning today. Whether that's talking to customers or talking to researchers at other labs, trying to understand both where the frontier is going and where things are really not falling apart today. And then developing a perspective on how we can push the frontier using interpretability methods. And so, you know, even our chief scientist, Tom, spends a lot of time talking to customers and trying to understand what real world problems are and then taking that back and trying to apply the current state of the art to those problems and then seeing where they fall down basically. And then using those failures or those shortcomings to understand what hills to climb when it comes to interpretability research. So like on the fundamental side, for instance, when we have done some work applying SAEs and probes, we've encountered, you know, some shortcomings in SAEs that we found a little bit surprising. And so have gone back to the drawing board and done work on that. And then, you know, we've done some work on better foundational interpreter models. And a lot of our team's research is focused on what is the next evolution beyond SAEs, for instance. And then when it comes to like control and design of models, you know, we tried steering with our first API and realized that it still fell short of black box techniques like prompting or fine tuning. And so went back to the drawing board and we're like, how do we make that not the case and how do we improve it beyond that? And one of our researchers, Ekdeep, who just joined is actually Ekdeep and Atticus are like steering experts and have spent a lot of time trying to figure out like, what is the research that enables us to actually do this in a much more powerful, robust way? So yeah, the answer is like, look at real world problems, try to translate that into a research agenda and then like hill climb on both of those at the same time.Shawn Wang [00:16:04]: Yeah. Mark has the steering CLI demo queued up, which we're going to go into in a sec. But I always want to double click on when you drop hints, like we found some problems with SAEs. Okay. What are they? You know, and then we can go into the demo. Yeah.Myra Deng [00:16:19]: I mean, I'm curious if you have more thoughts here as well, because you've done it in the healthcare domain. But I think like, for instance, when we do things like trying to detect behaviors within models that are harmful or like behaviors that a user might not want to have in their model. So hallucinations, for instance, harmful intent, PII, all of these things. We first tried using SAE probes for a lot of these tasks. So taking the feature activation space from SAEs and then training classifiers on top of that, and then seeing how well we can detect the properties that we might want to detect in model behavior. And we've seen in many cases that probes just trained on raw activations seem to perform better than SAE probes, which is a bit surprising if you think that SAEs are actually also capturing the concepts that you would want to capture cleanly and more surgically. And so that is an interesting observation. I don't think that is like, I'm not down on SAEs at all. I think there are many, many things they're useful for, but we have definitely run into cases where I think the concept space described by SAEs is not as clean and accurate as we would expect it to be for actual like real world downstream performance metrics.Mark Bissell [00:17:34]: Fair enough. Yeah. It's the blessing and the curse of unsupervised methods where you get to peek into the AI's mind. But sometimes you wish that you saw other things when you walked inside there. Although in the PII instance, I think weren't an SAE based approach actually did prove to be the most generalizable?Myra Deng [00:17:53]: It did work well in the case that we published with Rakuten. And I think a lot of the reasons it worked well was because we had a noisier data set. And so actually the blessing of unsupervised learning is that we actually got to get more meaningful, generalizable signal from SAEs when the data was noisy. But in other cases where we've had like good data sets, it hasn't been the case.Shawn Wang [00:18:14]: And just because you named Rakuten and I don't know if we'll get it another chance, like what is the overall, like what is Rakuten's usage or production usage? Yeah.Myra Deng [00:18:25]: So they are using us to essentially guardrail and inference time monitor their language model usage and their agent usage to detect things like PII so that they don't route private user information.Myra Deng [00:18:41]: And so that's, you know, going through all of their user queries every day. And that's something that we deployed with them a few months ago. And now we are actually exploring very early partnerships, not just with Rakuten, but with other people around how we can help with potentially training and customization use cases as well. Yeah.Shawn Wang [00:19:03]: And for those who don't know, like it's Rakuten is like, I think number one or number two e-commerce store in Japan. Yes. Yeah.Mark Bissell [00:19:10]: And I think that use case actually highlights a lot of like what it looks like to deploy things in practice that you don't always think about when you're doing sort of research tasks. So when you think about some of the stuff that came up there that's more complex than your idealized version of a problem, they were encountering things like synthetic to real transfer of methods. So they couldn't train probes, classifiers, things like that on actual customer data of PII. So what they had to do is use synthetic data sets. And then hope that that transfer is out of domain to real data sets. And so we can evaluate performance on the real data sets, but not train on customer PII. So that right off the bat is like a big challenge. You have multilingual requirements. So this needed to work for both English and Japanese text. Japanese text has all sorts of quirks, including tokenization behaviors that caused lots of bugs that caused us to be pulling our hair out. And then also a lot of tasks you'll see. You might make simplifying assumptions if you're sort of treating it as like the easiest version of the problem to just sort of get like general results where maybe you say you're classifying a sentence to say, does this contain PII? But the need that Rakuten had was token level classification so that you could precisely scrub out the PII. So as we learned more about the problem, you're sort of speaking about what that looks like in practice. Yeah. A lot of assumptions end up breaking. And that was just one instance where you. A problem that seems simple right off the bat ends up being more complex as you keep diving into it.Vibhu Sapra [00:20:41]: Excellent. One of the things that's also interesting with Interp is a lot of these methods are very efficient, right? So where you're just looking at a model's internals itself compared to a separate like guardrail, LLM as a judge, a separate model. One, you have to host it. Two, there's like a whole latency. So if you use like a big model, you have a second call. Some of the work around like self detection of hallucination, it's also deployed for efficiency, right? So if you have someone like Rakuten doing it in production live, you know, that's just another thing people should consider.Mark Bissell [00:21:12]: Yeah. And something like a probe is super lightweight. Yeah. It's no extra latency really. Excellent.Shawn Wang [00:21:17]: You have the steering demos lined up. So we were just kind of see what you got. I don't, I don't actually know if this is like the latest, latest or like alpha thing.Mark Bissell [00:21:26]: No, this is a pretty hacky demo from from a presentation that someone else on the team recently gave. So this will give a sense for, for technology. So you can see the steering and action. Honestly, I think the biggest thing that this highlights is that as we've been growing as a company and taking on kind of more and more ambitious versions of interpretability related problems, a lot of that comes to scaling up in various different forms. And so here you're going to see steering on a 1 trillion parameter model. This is Kimi K2. And so it's sort of fun that in addition to the research challenges, there are engineering challenges that we're now tackling. Cause for any of this to be sort of useful in production, you need to be thinking about what it looks like when you're using these methods on frontier models as opposed to sort of like toy kind of model organisms. So yeah, this was thrown together hastily, pretty fragile behind the scenes, but I think it's quite a fun demo. So screen sharing is on. So I've got two terminal sessions pulled up here. On the left is a forked version that we have of the Kimi CLI that we've got running to point at our custom hosted Kimi model. And then on the right is a set up that will allow us to steer on certain concepts. So I should be able to chat with Kimi over here. Tell it hello. This is running locally. So the CLI is running locally, but the Kimi server is running back to the office. Well, hopefully should be, um, that's too much to run on that Mac. Yeah. I think it's, uh, it takes a full, like each 100 node. I think it's like, you can. You can run it on eight GPUs, eight 100. So, so yeah, Kimi's running. We can ask it a prompt. It's got a forked version of our, uh, of the SG line code base that we've been working on. So I'm going to tell it, Hey, this SG line code base is slow. I think there's a bug. Can you try to figure it out? There's a big code base, so it'll, it'll spend some time doing this. And then on the right here, I'm going to initialize in real time. Some steering. Let's see here.Mark Bissell [00:23:33]: searching for any. Bugs. Feature ID 43205.Shawn Wang [00:23:38]: Yeah.Mark Bissell [00:23:38]: 20, 30, 40. So let me, uh, this is basically a feature that we found that inside Kimi seems to cause it to speak in Gen Z slang. And so on the left, it's still sort of thinking normally it might take, I don't know, 15 seconds for this to kick in, but then we're going to start hopefully seeing him do this code base is massive for real. So we're going to start. We're going to start seeing Kimi transition as the steering kicks in from normal Kimi to Gen Z Kimi and both in its chain of thought and its actual outputs.Mark Bissell [00:24:19]: And interestingly, you can see, you know, it's still able to call tools, uh, and stuff. It's um, it's purely sort of it's it's demeanor. And there are other features that we found for interesting things like concision. So that's more of a practical one. You can make it more concise. Um, the types of programs, uh, programming languages that uses, but yeah, as we're seeing it come in. Pretty good. Outputs.Shawn Wang [00:24:43]: Scheduler code is actually wild.Vibhu Sapra [00:24:46]: Yo, this code is actually insane, bro.Vibhu Sapra [00:24:53]: What's the process of training in SAE on this, or, you know, how do you label features? I know you guys put out a pretty cool blog post about, um, finding this like autonomous interp. Um, something. Something about how agents for interp is different than like coding agents. I don't know while this is spewing up, but how, how do we find feature 43, two Oh five. Yeah.Mark Bissell [00:25:15]: So in this case, um, we, our platform that we've been building out for a long time now supports all the sort of classic out of the box interp techniques that you might want to have like SAE training, probing things of that kind, I'd say the techniques for like vanilla SAEs are pretty well established now where. You take your model that you're interpreting, run a whole bunch of data through it, gather activations, and then yeah, pretty straightforward pipeline to train an SAE. There are a lot of different varieties. There's top KSAEs, batch top KSAEs, um, normal ReLU SAEs. And then once you have your sparse features to your point, assigning labels to them to actually understand that this is a gen Z feature, that's actually where a lot of the kind of magic happens. Yeah. And the most basic standard technique is look at all of your d input data set examples that cause this feature to fire most highly. And then you can usually pick out a pattern. So for this feature, If I've run a diverse enough data set through my model feature 43, two Oh five. Probably tends to fire on all the tokens that sounds like gen Z slang. You know, that's the, that's the time of year to be like, Oh, I'm in this, I'm in this Um, and, um, so, you know, you could have a human go through all 43,000 concepts andVibhu Sapra [00:26:34]: And I've got to ask the basic question, you know, can we get examples where it hallucinates, pass it through, see what feature activates for hallucinations? Can I just, you know, turn hallucination down?Myra Deng [00:26:51]: Oh, wow. You really predicted a project we're already working on right now, which is detecting hallucinations using interpretability techniques. And this is interesting because hallucinations is something that's very hard to detect. And it's like a kind of a hairy problem and something that black box methods really struggle with. Whereas like Gen Z, you could always train a simple classifier to detect that hallucinations is harder. But we've seen that models internally have some... Awareness of like uncertainty or some sort of like user pleasing behavior that leads to hallucinatory behavior. And so, yeah, we have a project that's trying to detect that accurately. And then also working on mitigating the hallucinatory behavior in the model itself as well.Shawn Wang [00:27:39]: Yeah, I would say most people are still at the level of like, oh, I would just turn temperature to zero and that turns off hallucination. And I'm like, well, that's a fundamental misunderstanding of how this works. Yeah.Mark Bissell [00:27:51]: Although, so part of what I like about that question is you, there are SAE based approaches that might like help you get at that. But oftentimes the beauty of SAEs and like we said, the curse is that they're unsupervised. So when you have a behavior that you deliberately would like to remove, and that's more of like a supervised task, often it is better to use something like probes and specifically target the thing that you're interested in reducing as opposed to sort of like hoping that when you fragment the latent space, one of the vectors that pops out.Vibhu Sapra [00:28:20]: And as much as we're training an autoencoder to be sparse, we're not like for sure certain that, you know, we will get something that just correlates to hallucination. You'll probably split that up into 20 other things and who knows what they'll be.Mark Bissell [00:28:36]: Of course. Right. Yeah. So there's no sort of problems with like feature splitting and feature absorption. And then there's the off target effects, right? Ideally, you would want to be very precise where if you reduce the hallucination feature, suddenly maybe your model can't write. Creatively anymore. And maybe you don't like that, but you want to still stop it from hallucinating facts and figures.Shawn Wang [00:28:55]: Good. So Vibhu has a paper to recommend there that we'll put in the show notes. But yeah, I mean, I guess just because your demo is done, any any other things that you want to highlight or any other interesting features you want to show?Mark Bissell [00:29:07]: I don't think so. Yeah. Like I said, this is a pretty small snippet. I think the main sort of point here that I think is exciting is that there's not a whole lot of inter being applied to models quite at this scale. You know, Anthropic certainly has some some. Research and yeah, other other teams as well. But it's it's nice to see these techniques, you know, being put into practice. I think not that long ago, the idea of real time steering of a trillion parameter model would have sounded.Shawn Wang [00:29:33]: Yeah. The fact that it's real time, like you started the thing and then you edited the steering vector.Vibhu Sapra [00:29:38]: I think it's it's an interesting one TBD of what the actual like production use case would be on that, like the real time editing. It's like that's the fun part of the demo, right? You can kind of see how this could be served behind an API, right? Like, yes, you're you only have so many knobs and you can just tweak it a bit more. And I don't know how it plays in. Like people haven't done that much with like, how does this work with or without prompting? Right. How does this work with fine tuning? Like, there's a whole hype of continual learning, right? So there's just so much to see. Like, is this another parameter? Like, is it like parameter? We just kind of leave it as a default. We don't use it. So I don't know. Maybe someone here wants to put out a guide on like how to use this with prompting when to do what?Mark Bissell [00:30:18]: Oh, well, I have a paper recommendation. I think you would love from Act Deep on our team, who is an amazing researcher, just can't say enough amazing things about Act Deep. But he actually has a paper that as well as some others from the team and elsewhere that go into the essentially equivalence of activation steering and in context learning and how those are from a he thinks of everything in a cognitive neuroscience Bayesian framework, but basically how you can precisely show how. Prompting in context, learning and steering exhibit similar behaviors and even like get quantitative about the like magnitude of steering you would need to do to induce a certain amount of behavior similar to certain prompting, even for things like jailbreaks and stuff. It's a really cool paper. Are you saying steering is less powerful than prompting? More like you can almost write a formula that tells you how to convert between the two of them.Myra Deng [00:31:20]: And so like formally equivalent actually in the in the limit. Right.Mark Bissell [00:31:24]: So like one case study of this is for jailbreaks there. I don't know. Have you seen the stuff where you can do like many shot jailbreaking? You like flood the context with examples of the behavior. And the topic put out that paper.Shawn Wang [00:31:38]: A lot of people were like, yeah, we've been doing this, guys.Mark Bissell [00:31:40]: Like, yeah, what's in this in context learning and activation steering equivalence paper is you can like predict the number. Number of examples that you will need to put in there in order to jailbreak the model. That's cool. By doing steering experiments and using this sort of like equivalence mapping. That's cool. That's really cool. It's very neat. Yeah.Shawn Wang [00:32:02]: I was going to say, like, you know, I can like back rationalize that this makes sense because, you know, what context is, is basically just, you know, it updates the KV cache kind of and like and then every next token inference is still like, you know, the sheer sum of everything all the way. It's plus all the context. It's up to date. And you could, I guess, theoretically steer that with you probably replace that with your steering. The only problem is steering typically is on one layer, maybe three layers like like you did. So it's like not exactly equivalent.Mark Bissell [00:32:33]: Right, right. There's sort of you need to get precise about, yeah, like how you sort of define steering and like what how you're modeling the setup. But yeah, I've got the paper pulled up here. Belief dynamics reveal the dual nature. Yeah. The title is Belief Dynamics Reveal the Dual Nature of Incompetence. And it's an exhibition of the practical context learning and activation steering. So Eric Bigelow, Dan Urgraft on the who are doing fellowships at Goodfire, Ekt Deep's the final author there.Myra Deng [00:32:59]: I think actually to your question of like, what is the production use case of steering? I think maybe if you just think like one level beyond steering as it is today. Like imagine if you could adapt your model to be, you know, an expert legal reasoner. Like in almost real time, like very quickly. efficiently using human feedback or using like your semantic understanding of what the model knows and where it knows that behavior. I think that while it's not clear what the product is at the end of the day, it's clearly very valuable. Thinking about like what's the next interface for model customization and adaptation is a really interesting problem for us. Like we have heard a lot of people actually interested in fine-tuning an RL for open weight models in production. And so people are using things like Tinker or kind of like open source libraries to do that, but it's still very difficult to get models fine-tuned and RL'd for exactly what you want them to do unless you're an expert at model training. And so that's like something we'reShawn Wang [00:34:06]: looking into. Yeah. I never thought so. Tinker from Thinking Machines famously uses rank one LoRa. Is that basically the same as steering? Like, you know, what's the comparison there?Mark Bissell [00:34:19]: Well, so in that case, you are still applying updates to the parameters, right?Shawn Wang [00:34:25]: Yeah. You're not touching a base model. You're touching an adapter. It's kind of, yeah.Mark Bissell [00:34:30]: Right. But I guess it still is like more in parameter space then. I guess it's maybe like, are you modifying the pipes or are you modifying the water flowing through the pipes to get what you're after? Yeah. Just maybe one way.Mark Bissell [00:34:44]: I like that analogy. That's my mental map of it at least, but it gets at this idea of model design and intentional design, which is something that we're, that we're very focused on. And just the fact that like, I hope that we look back at how we're currently training models and post-training models and just think what a primitive way of doing that right now. Like there's no intentionalityShawn Wang [00:35:06]: really in... It's just data, right? The only thing in control is what data we feed in.Mark Bissell [00:35:11]: So, so Dan from Goodfire likes to use this analogy of, you know, he has a couple of young kids and he talks about like, what if I could only teach my kids how to be good people by giving them cookies or like, you know, giving them a slap on the wrist if they do something wrong, like not telling them why it was wrong or like what they should have done differently or something like that. Just figure it out. Right. Exactly. So that's RL. Yeah. Right. And, and, you know, it's sample inefficient. There's, you know, what do they say? It's like slurping feedback. It's like, slurping supervision. Right. And so you'd like to get to the point where you can have experts giving feedback to their models that are, uh, internalized and, and, you know, steering is an inference time way of sort of getting that idea. But ideally you're moving to a world whereVibhu Sapra [00:36:04]: it is much more intentional design in perpetuity for these models. Okay. This is one of the questions we asked Emmanuel from Anthropic on the podcast a few months ago. Basically the question, was you're at a research lab that does model training, foundation models, and you're on an interp team. How does it tie back? Right? Like, does this, do ideas come from the pre-training team? Do they go back? Um, you know, so for those interested, you can, you can watch that. There wasn't too much of a connect there, but it's still something, you know, it's something they want toMark Bissell [00:36:33]: push for down the line. It can be useful for all of the above. Like there are certainly post-hocVibhu Sapra [00:36:39]: use cases where it doesn't need to touch that. I think the other thing a lot of people forget is this stuff isn't too computationally expensive, right? Like I would say, if you're interested in getting into research, MechInterp is one of the most approachable fields, right? A lot of this train an essay, train a probe, this stuff, like the budget for this one, there's already a lot done. There's a lot of open source work. You guys have done some too. Um, you know,Shawn Wang [00:37:04]: There's like notebooks from the Gemini team for Neil Nanda or like, this is how you do it. Just step through the notebook.Vibhu Sapra [00:37:09]: Even if you're like, not even technical with any of this, you can still make like progress. There, you can look at different activations, but, uh, if you do want to get into training, you know, training this stuff, correct me if I'm wrong is like in the thousands of dollars, not even like, it's not that high scale. And then same with like, you know, applying it, doing it for post-training or all this stuff is fairly cheap in scale of, okay. I want to get into like model training. I don't have compute for like, you know, pre-training stuff. So it's, it's a very nice field to get into. And also there's a lot of like open questions, right? Um, some of them have to go with, okay, I want a product. I want to solve this. Like there's also just a lot of open-ended stuff that people could work on. That's interesting. Right. I don't know if you guys have any calls for like, what's open questions, what's open work that you either open collaboration with, or like, you'd just like to see solved or just, you know, for people listening that want to get into McInturk because people always talk about it. What are, what are the things they should check out? Start, of course, you know, join you guys as well. I'm sure you're hiring.Myra Deng [00:38:09]: There's a paper, I think from, was it Lee, uh, Sharky? It's open problems and, uh, it's, it's a bit of interpretability, which I recommend everyone who's interested in the field. Read. I'm just like a really comprehensive overview of what are the things that experts in the field think are the most important problems to be solved. I also think to your point, it's been really, really inspiring to see, I think a lot of young people getting interested in interpretability, actually not just young people also like scientists to have been, you know, experts in physics for many years and in biology or things like this, um, transitioning into interp, because the barrier of, of what's now interp. So it's really cool to see a number to entry is, you know, in some ways low and there's a lot of information out there and ways to get started. There's this anecdote of like professors at universities saying that all of a sudden every incoming PhD student wants to study interpretability, which was not the case a few years ago. So it just goes to show how, I guess, like exciting the field is, how fast it's moving, how quick it is to get started and things like that.Mark Bissell [00:39:10]: And also just a very welcoming community. You know, there's an open source McInturk Slack channel. There are people are always posting questions and just folks in the space are always responsive if you ask things on various forums and stuff. But yeah, the open paper, open problems paper is a really good one.Myra Deng [00:39:28]: For other people who want to get started, I think, you know, MATS is a great program. What's the acronym for? Machine Learning and Alignment Theory Scholars? It's like the...Vibhu Sapra [00:39:40]: Normally summer internship style.Myra Deng [00:39:42]: Yeah, but they've been doing it year round now. And actually a lot of our full-time staff have come through that program or gone through that program. And it's great for anyone who is transitioning into interpretability. There's a couple other fellows programs. We do one as well as Anthropic. And so those are great places to get started if anyone is interested.Mark Bissell [00:40:03]: Also, I think been seen as a research field for a very long time. But I think engineering... I think engineers are sorely wanted for interpretability as well, especially at Goodfire, but elsewhere, as it does scale up.Shawn Wang [00:40:18]: I should mention that Lee actually works with you guys, right? And in the London office and I'm adding our first ever McInturk track at AI Europe because I see this industry applications now emerging. And I'm pretty excited to, you know, help push that along. Yeah, I was looking forward to that. It'll effectively be the first industry McInturk conference. Yeah. I'm so glad you added that. You know, it's still a little bit of a bet. It's not that widespread, but I can definitely see this is the time to really get into it. We want to be early on things.Mark Bissell [00:40:51]: For sure. And I think the field understands this, right? So at ICML, I think the title of the McInturk workshop this year was actionable interpretability. And there was a lot of discussion around bringing it to various domains. Everyone's adding pragmatic, actionable, whatever.Shawn Wang [00:41:10]: It's like, okay, well, we weren't actionable before, I guess. I don't know.Vibhu Sapra [00:41:13]: And I mean, like, just, you know, being in Europe, you see the Interp room. One, like old school conferences, like, I think they had a very tiny room till they got lucky and they got it doubled. But there's definitely a lot of interest, a lot of niche research. So you see a lot of research coming out of universities, students. We covered the paper last week. It's like two unknown authors, not many citations. But, you know, you can make a lot of meaningful work there. Yeah. Yeah. Yeah.Shawn Wang [00:41:39]: Yeah. I think people haven't really mentioned this yet. It's just Interp for code. I think it's like an abnormally important field. We haven't mentioned this yet. The conspiracy theory last two years ago was when the first SAE work came out of Anthropic was they would do like, oh, we just used SAEs to turn the bad code vector down and then turn up the good code. And I think like, isn't that the dream? Like, you know, like, but basically, I guess maybe, why is it funny? Like, it's... If it was realistic, it would not be funny. It would be like, no, actually, we should do this. But it's funny because we know there's like, we feel there's some limitations to what steering can do. And I think a lot of the public image of steering is like the Gen Z stuff. Like, oh, you can make it really love the Golden Gate Bridge, or you can make it speak like Gen Z. To like be a legal reasoner seems like a huge stretch. Yeah. And I don't know if that will get there this way. Yeah.Myra Deng [00:42:36]: I think, um, I will say we are announcing. Something very soon that I will not speak too much about. Um, but I think, yeah, this is like what we've run into again and again is like, we, we don't want to be in the world where steering is only useful for like stylistic things. That's definitely not, not what we're aiming for. But I think the types of interventions that you need to do to get to things like legal reasoning, um, are much more sophisticated and require breakthroughs in, in learning algorithms. And that's, um...Shawn Wang [00:43:07]: And is this an emergent property of scale as well?Myra Deng [00:43:10]: I think so. Yeah. I mean, I think scale definitely helps. I think scale allows you to learn a lot of information and, and reduce noise across, you know, large amounts of data. But I also think we think that there's ways to do things much more effectively, um, even, even at scale. So like actually learning exactly what you want from the data and not learning things that you do that you don't want exhibited in the data. So we're not like anti-scale, but we are also realizing that scale is not going to get us anywhere. It's not going to get us to the type of AI development that we want to be at in, in the future as these models get more powerful and get deployed in all these sorts of like mission critical contexts. Current life cycle of training and deploying and evaluations is, is to us like deeply broken and has opportunities to, to improve. So, um, more to come on that very, very soon.Mark Bissell [00:44:02]: And I think that that's a use basically, or maybe just like a proof point that these concepts do exist. Like if you can manipulate them in the precise best way, you can get the ideal combination of them that you desire. And steering is maybe the most coarse grained sort of peek at what that looks like. But I think it's evocative of what you could do if you had total surgical control over every concept, every parameter. Yeah, exactly.Myra Deng [00:44:30]: There were like bad code features. I've got it pulled up.Vibhu Sapra [00:44:33]: Yeah. Just coincidentally, as you guys are talking.Shawn Wang [00:44:35]: This is like, this is exactly.Vibhu Sapra [00:44:38]: There's like specifically a code error feature that activates and they show, you know, it's not, it's not typo detection. It's like, it's, it's typos in code. It's not typical typos. And, you know, you can, you can see it clearly activates where there's something wrong in code. And they have like malicious code, code error. They have a whole bunch of sub, you know, sub broken down little grain features. Yeah.Shawn Wang [00:45:02]: Yeah. So, so the, the rough intuition for me, the, why I talked about post-training was that, well, you just, you know, have a few different rollouts with all these things turned off and on and whatever. And then, you know, you can, that's, that's synthetic data you can kind of post-train on. Yeah.Vibhu Sapra [00:45:13]: And I think we make it sound easier than it is just saying, you know, they do the real hard work.Myra Deng [00:45:19]: I mean, you guys, you guys have the right idea. Exactly. Yeah. We replicated a lot of these features in, in our Lama models as well. I remember there was like.Vibhu Sapra [00:45:26]: And I think a lot of this stuff is open, right? Like, yeah, you guys opened yours. DeepMind has opened a lot of essays on Gemma. Even Anthropic has opened a lot of this. There's, there's a lot of resources that, you know, we can probably share of people that want to get involved.Shawn Wang [00:45:41]: Yeah. And special shout out to like Neuronpedia as well. Yes. Like, yeah, amazing piece of work to visualize those things.Myra Deng [00:45:49]: Yeah, exactly.Shawn Wang [00:45:50]: I guess I wanted to pivot a little bit on, onto the healthcare side, because I think that's a big use case for you guys. We haven't really talked about it yet. This is a bit of a crossover for me because we are, we are, we do have a separate science pod that we're starting up for AI, for AI for science, just because like, it's such a huge investment category and also I'm like less qualified to do it, but we actually have bio PhDs to cover that, which is great, but I need to just kind of recover, recap your work, maybe on the evil two stuff, but then, and then building forward.Mark Bissell [00:46:17]: Yeah, for sure. And maybe to frame up the conversation, I think another kind of interesting just lens on interpretability in general is a lot of the techniques that were described. are ways to solve the AI human interface problem. And it's sort of like bidirectional communication is the goal there. So what we've been talking about with intentional design of models and, you know, steering, but also more advanced techniques is having humans impart our desires and control into models and over models. And the reverse is also very interesting, especially as you get to superhuman models, whether that's narrow superintelligence, like these scientific models that work on genomics, data, medical imaging, things like that. But down the line, you know, superintelligence of other forms as well. What knowledge can the AIs teach us as sort of that, that the other direction in that? And so some of our life science work to date has been getting at exactly that question, which is, well, some of it does look like debugging these various life sciences models, understanding if they're actually performing well, on tasks, or if they're picking up on spurious correlations, for instance, genomics models, you would like to know whether they are sort of focusing on the biologically relevant things that you care about, or if it's using some simpler correlate, like the ancestry of the person that it's looking at. But then also in the instances where they are superhuman, and maybe they are understanding elements of the human genome that we don't have names for or specific, you know, yeah, discoveries that they've made that that we don't know about, that's, that's a big goal. And so we're already seeing that, right, we are partnered with organizations like Mayo Clinic, leading research health system in the United States, our Institute, as well as a startup called Prima Menta, which focuses on neurodegenerative disease. And in our partnership with them, we've used foundation models, they've been training and applied our interpretability techniques to find novel biomarkers for Alzheimer's disease. So I think this is just the tip of the iceberg. But it's, that's like a flavor of some of the things that we're working on.Shawn Wang [00:48:36]: Yeah, I think that's really fantastic. Obviously, we did the Chad Zuckerberg pod last year as well. And like, there's a plethora of these models coming out, because there's so much potential and research. And it's like, very interesting how it's basically the same as language models, but just with a different underlying data set. But it's like, it's the same exact techniques. Like, there's no change, basically.Mark Bissell [00:48:59]: Yeah. Well, and even in like other domains, right? Like, you know, robotics, I know, like a lot of the companies just use Gemma as like the like backbone, and then they like make it into a VLA that like takes these actions. It's, it's, it's transformers all the way down. So yeah.Vibhu Sapra [00:49:15]: Like we have Med Gemma now, right? Like this week, even there was Med Gemma 1.5. And they're training it on this stuff, like 3d scans, medical domain knowledge, and all that stuff, too. So there's a push from both sides. But I think the thing that, you know, one of the things about McInturpp is like, you're a little bit more cautious in some domains, right? So healthcare, mainly being one, like guardrails, understanding, you know, we're more risk adverse to something going wrong there. So even just from a basic understanding, like, if we're trusting these systems to make claims, we want to know why and what's going on.Myra Deng [00:49:51]: Yeah, I think there's totally a kind of like deployment bottleneck to actually using. foundation models for real patient usage or things like that. Like, say you're using a model for rare disease prediction, you probably want some explanation as to why your model predicted a certain outcome, and an interpretable explanation at that. So that's definitely a use case. But I also think like, being able to extract scientific information that no human knows to accelerate drug discovery and disease treatment and things like that actually is a really, really big unlock for science, like scientific discovery. And you've seen a lot of startups, like say that they're going to accelerate scientific discovery. And I feel like we actually are doing that through our interp techniques. And kind of like, almost by accident, like, I think we got reached out to very, very early on from these healthcare institutions. And none of us had healthcare.Shawn Wang [00:50:49]: How did they even hear of you? A podcast.Myra Deng [00:50:51]: Oh, okay. Yeah, podcast.Vibhu Sapra [00:50:53]: Okay, well, now's that time, you know.Myra Deng [00:50:55]: Everyone can call us.Shawn Wang [00:50:56]: Podcasts are the most important thing. Everyone should listen to podcasts.Myra Deng [00:50:59]: Yeah, they reached out. They were like, you know, we have these really smart models that we've trained, and we want to know what they're doing. And we were like, really early that time, like three months old, and it was a few of us. And we were like, oh, my God, we've never used these models. Let's figure it out. But it's also like, great proof that interp techniques scale pretty well across domains. We didn't really have to learn too much about.Shawn Wang [00:51:21]: Interp is a machine learning technique, machine learning skills everywhere, right? Yeah. And it's obviously, it's just like a general insight. Yeah. Probably to finance too, I think, which would be fun for our history. I don't know if you have anything to say there.Mark Bissell [00:51:34]: Yeah, well, just across the science. Like, we've also done work on material science. Yeah, it really runs the gamut.Vibhu Sapra [00:51:40]: Yeah. Awesome. And, you know, for those that should reach out, like, you're obviously experts in this, but like, is there a call out for people that you're looking to partner with, design partners, people to use your stuff outside of just, you know, the general developer that wants to. Plug and play steering stuff, like on the research side more so, like, are there ideal design partners, customers, stuff like that?Myra Deng [00:52:03]: Yeah, I can talk about maybe non-life sciences, and then I'm curious to hear from you on the life sciences side. But we're looking for design partners across many domains, language, anyone who's customizing language models or trying to push the frontier of code or reasoning models is really interesting to us. And then also interested in the frontier of modeling. There's a lot of models that work in, like, pixel space, as we call it. So if you're doing world models, video models, even robotics, where there's not a very clean natural language interface to interact with, I think we think that Interp can really help and are looking for a few partners in that space.Shawn Wang [00:52:43]: Just because you mentioned the keyword

Unbiased Science
No Representation, No Generalization: Health Equity in Research

Unbiased Science

Play Episode Listen Later Feb 4, 2026 34:13


In this episode, Jess and Sarah welcome Dr. Kate Wallis and Dr. Diana Montoya-Williams to explore the essential topic of health equity in scientific research. The scientists examine the critical importance of rigorous research design and the transformative role of community engagement in conducting meaningful health studies. They address common methodological mistakes that compromise research validity, particularly focusing on how race and ethnicity are contextualized in scientific studies. Throughout the conversation, there is an emphasis on the need for greater transparency in research practices and how community involvement strengthens both the quality and relevance of scientific work. Despite acknowledging significant challenges in achieving health equity, the episode concludes on a hopeful note by highlighting the power of community solidarity and engagement in advancing public health outcomes. Watch the conversation on YouTube: https://youtu.be/p726HlABGRI (00:00) Intro & Public Health Update (04:22) What's A Health/Science News Item That Caught Your Attention? (06:43) A Collaborative Project About How Science Has Failed Certain Communities (12:04) Common Mistakes In Research Validity (16:24) Understanding Race & Ethnicity In Research (21:25) What Does True Community Engagement Look Like? (30:07) What's Giving You Hope In Public Health And Science Right Now? https://www.inquirer.com/health/expert-opinions/autism-treatments-myths-fda-cdc-changes-20251204.html https://publications.aap.org/pediatricsopenscience/article/2/1/1/205504/Consensus-Recommendations-for-Antiracist-Child?searchresult=1 ----------------------------------------------------------------------------------------------------------------------- Interested in advertising with us? Please reach out to advertising@airwavemedia.com, with “Unbiased Science” in the subject line. PLEASE NOTE: The discussion and information provided in this podcast are for general educational, scientific, and informational purposes only and are not intended as, and should not be treated as, medical or other professional advice for any particular individual or individuals. Every person and medical issue is different, and diagnosis and treatment requires consideration of specific facts often unique to the individual. As such, the information contained in this podcast should not be used as a substitute for consultation with and/or treatment by a doctor or other medical professional. If you are experiencing any medical issue or have any medical concern, you should consult with a doctor or other medical professional. Further, due to the inherent limitations of a podcast such as this as well as ongoing scientific developments, we do not guarantee the completeness or accuracy of the information or analysis provided in this podcast, although, of course we always endeavor to provide comprehensive information and analysis. In no event may Unbiased Science or any of the participants in this podcast be held liable to the listener or anyone else for any decision allegedly made or action allegedly taken or not taken allegedly in reliance on the discussion or information in this podcast or for any damages allegedly resulting from such reliance. The information provided herein do not represent the views of our employers. Learn more about your ad choices. Visit megaphone.fm/adchoices

research representation health equity generalization consensus recommendations unbiased science
Unsupervised Learning
Ep 81: Ex-OpenAI Researcher On Why He Left, His Honest AGI Timeline, & The Limits of Scaling RL

Unsupervised Learning

Play Episode Listen Later Jan 29, 2026 62:52


This episode features Jerry Tworek, a key architect behind OpenAI's breakthrough reasoning models (o1, o3) and Codex, discussing the current state and future of AI. Jerry explores the real limits and promise of scaling pre-training and reinforcement learning, arguing that while these paradigms deliver predictable improvements, they're fundamentally constrained by data availability and struggle with generalization beyond their training objectives. He reveals his updated belief that continual learning—the ability for models to update themselves based on failure and work through problems autonomously—is necessary for AGI, as current models hit walls and become "hopeless" when stuck. Jerry discusses the convergence of major labs toward similar approaches driven by economic forces, the tension between exploration and exploitation in research, and why he left OpenAI to pursue new research directions. He offers candid insights on the competitive dynamics between labs, the focus required to win in specific domains like coding, what makes great AI researchers, and his surprisingly near-term predictions for robotics (2-3 years) while warning about the societal implications of widespread work automation that we're not adequately preparing for. (0:00) Intro(1:26) Scaling Paradigms in AI(3:36) Challenges in Reinforcement Learning(11:48) AGI Timelines(18:36) Converging Labs(25:05) Jerry's Departure from OpenAI(31:18) Pivotal Decisions in OpenAI's Journey(35:06) Balancing Research and Product Development(38:42) The Future of AI Coding(41:33) Specialization vs. Generalization in AI(48:47) Hiring and Building Research Teams(55:21) Quickfire With your co-hosts: @jacobeffron - Partner at Redpoint, Former PM Flatiron Health @patrickachase - Partner at Redpoint, Former ML Engineer LinkedIn @ericabrescia - Former COO Github, Founder Bitnami (acq'd by VMWare) @jordan_segall - Partner at Redpoint

Canine Revolution Radio
#148 The Ultimate Leash Training Manual | 5 Steps to a Well Behaved Dog

Canine Revolution Radio

Play Episode Listen Later Jan 12, 2026 43:22


You've probably tried every trick in the book to stop your dog from pulling on the leash... Maybe you've switched collars three times, bought a "no-pull" harness, or even considered giving up on walks altogether...making you feel like you're the only owner who can't control their dog on a simple walk. However, that's not the case at all.We've developed a proven 5-phase system that transforms even the worst leash pullers into well-behaved walking partners.Today we're breaking down the exact progression from my new book "The Ultimate Leash Training Manual" that's helped thousands of dogs master loose leash walking.Recommended Training Equipment:

Huberman Lab
Best Ways to Build Better Habits & Break Bad Ones | James Clear

Huberman Lab

Play Episode Listen Later Jan 5, 2026 155:33


James Clear is an expert on behavioral change and habits and the author of the bestselling book Atomic Habits. We discuss the best ways to build new healthy habits and end bad ones without relying on motivation or willpower. Rather than list off categories of tools or acronyms, James explains how anchoring the changes you want to make in your identity and physical environment allows you to make desired changes quickly and ones that stick. Whether your goal is better fitness and physical health, productivity or mental health, you'll learn actionable, zero-cost protocols to build powerful and meaningful habits. Sponsors AG1: https://drinkag1.com/huberman Lingo: https://hellolingo.com/huberman Wealthfront*: https://wealthfront.com/huberman Joovv: https://joovv.com/huberman Eight Sleep: https://eightsleep.com/huberman Function: https://functionhealth.com/huberman Timestamps 00:00:00 James Clear 00:02:57 Common Habits, Tool: Habit Success & Getting Started 00:06:16 Make Starting a Habit Easier, Tool: 4 Laws of Behavior Change 00:10:18 Sponsors: Lingo & Wealthfront 00:13:26 Writing Habits, Seasons & Flexibility; Adaptability, Tool: Bad Day Plan 00:18:42 Consistency, Flow vs Grind, Master Showing Up, Learning & Practice 00:24:54 Chunking, Getting Started at Gym 00:28:01 Flow Don't Fight, Dissatisfaction & Effort, Tool: Identity-Based Habits 00:34:10 Friction, Competition & Effort; Credentials 00:39:38 Make Effort Rewarding, Mindset, Tools: Previsualization, Emphasize Positives 00:45:59 Sponsors: AG1 & Joovv 00:48:56 Reflection & Learning, Tool: Self-Testing; Perfectionism, Tool: Curiosity 00:55:18 Striving vs Relaxation, Balance, Tool: Turn On/Off; Hiking, Nature Reset 01:04:20 Identity & Professional Pursuits; Choosing New Projects; Clinging to Identity 01:14:24 Sponsor: Eight Sleep 01:15:42 Criticism; Identity & Growth 01:21:47 Failure, Identity, Sports, Tool: Rebounding & Reaching; Public Failures 01:30:03 Daily Habits, Tools: Day in Quarters; Never Miss Twice; Meal Timing 01:38:22 Daily Habit Timing & Sequencing, Tool: Mindfully Choose Inputs 01:45:37 Creativity, Specialization vs Generalization; Books 01:51:31 Sponsor: Function 01:53:18 Habits & Context, Environmental Cues, Tools for Minimizing Phone Use 02:02:01 Bad Habits, Checking Phone, Tools for Breaking Bad Habits 02:08:21 Physical & Social Environment, New Habits, Tool: Join/Create Groups 02:18:40 Family, Habits; Kids & Parenting, Tools: Stimulus; Good Conditions 02:26:05 Impact of Habits, Habits as Solutions; Upcoming Projects 02:32:45 Zero-Cost Support, YouTube, Spotify & Apple Follow, Reviews & Feedback, Sponsors, Protocols Book, Social Media, Neural Network Newsletter *This experience may not be representative of other Wealthfront clients, and there is no guarantee of future performance or success. Experiences will vary. The Cash Account, which is not a deposit account, is offered by Wealthfront Brokerage LLC, member FINRA/SIPC.  Wealthfront Brokerage is not a bank. The base APY is 3.50% on cash deposits as of November 07, 2025, is representative, subject to change, and requires no minimum. If eligible for the overall boosted rate of 4.15% offered in connection with this promo, your boosted rate is also subject to change if the base rate decreases during the 3 month promo period. Funds in the Cash Account are swept to program banks, where it earns the variable APY. New Cash Account deposits are subject to a 2-4 day holding period before becoming available for transfer. Investment advisory services are provided by Wealthfront Advisers LLC, an SEC-registered investment adviser. Securities investments: not bank deposits, bank-guaranteed or FDIC-insured, and may lose value. Learn more about your ad choices. Visit megaphone.fm/adchoices

Enrichment for the Real World
#147 - Old Skill, New Scenario: Using What You Already Know

Enrichment for the Real World

Play Episode Listen Later Dec 29, 2025 48:17


If your first response to a new behavior challenge is “I need to learn something new,” this episode is for you. Ellen and Emily talk about why “new” isn't always the answer, and how to make the most of the skills already in your toolbox. From spooky sedation stories to “my perfect puppy isn't perfect anymore” meltdowns, they'll help you see that the solutions you need might already be sitting right there, waiting to be dusted off.TLDL (too long, didn't listen): 3 Key Takeaways 1️⃣Solid foundations beat shiny new tools - The basics, done well, solve more problems than you'd think.2️⃣Generalization is underrated - The real magic happens when you and your pet can use familiar tools in new ways.3️⃣ You're not starting over. You're leveling up - Every challenge is just another chance to practice what you already know.For the full episode show notes, including the resources mentioned in this episode, go here.More from Pet HarmonyPet Parents: enrichment ideas and practical behavior tips

Let's Talk AI
#228 - GPT 5.2, Scaling Agents, Weird Generalization

Let's Talk AI

Play Episode Listen Later Dec 17, 2025 86:42


Our 228th episode with a summary and discussion of last week's big AI news!Recorded on 12/12/2025Hosted by Andrey Kurenkov and Jeremie HarrisFeel free to email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.aiRead out our text newsletter and comment on the podcast at https://lastweekin.ai/In this episode:OpenAI's latest model GPT-5.2 demonstrates improved performance and enhanced multi-modal capabilities but comes with increased costs and a different knowledge cutoff date.Disney invests $1 billion in OpenAI to generate Disney character content, creating unique licensing agreements across characters from Marvel, Pixar, and Star Wars franchises.The U.S. government imposes new AI chip export rules involving security reviews, while simultaneously moving to prevent states from independently regulating AI.DeepMind releases a paper outlining the challenges and findings in scaling multi-agent systems, highlighting the complexities of tool coordination and task performance.Timestamps:(00:00:00) Intro / Banter(00:01:19) News PreviewTools & Apps(00:01:58) GPT-5.2 is OpenAI's latest move in the agentic AI battle | The Verge(00:08:48) Runway releases its first world model, adds native audio to latest video model | TechCrunch(00:11:51) Google says it will link to more sources in AI Mode | The Verge(00:12:24) ChatGPT can now use Adobe apps to edit your photos and PDFs for free | The Verge(00:13:05) Tencent releases Hunyuan 2.0 with 406B parametersApplications & Business(00:16:15) China set to limit access to Nvidia's H200 chips despite Trump export approval(00:21:02) Disney investing $1 billion in OpenAI, will allow characters on Sora(00:24:48) Unconventional AI confirms its massive $475M seed round(00:29:06) Slack CEO Denise Dresser to join OpenAI as chief revenue officer | TechCrunch(00:31:18) The state of enterprise AIProjects & Open Source(00:33:49) [2512.10791] The FACTS Leaderboard: A Comprehensive Benchmark for Large Language Model Factuality(00:36:27) Claude 4.5 Opus' Soul DocumentResearch & Advancements(00:43:49) [2512.08296] Towards a Science of Scaling Agent Systems(00:48:43) Evaluating Gemini Robotics Policies in a Veo World Simulator(00:52:10) Guided Self-Evolving LLMs with Minimal Human Supervision(00:56:08) Martingale Score: An Unsupervised Metric for Bayesian Rationality in LLM Reasoning(01:00:39) [2512.07783] On the Interplay of Pre-Training, Mid-Training, and RL on Reasoning Language Models(01:04:42) Stabilizing Reinforcement Learning with LLMs: Formulation and Practices(01:09:42) Google's AI unit DeepMind announces UK 'automated research lab'Policy & Safety(01:10:28) Trump Moves to Stop States From Regulating AI With a New Executive Order - The New York Times(01:13:54) [2512.09742] Weird Generalization and Inductive Backdoors: New Ways to Corrupt LLMs(01:17:57) Forecasting AI Time Horizon Under Compute Slowdowns(01:20:46) AI Security Institute focuses on AI measurements and evaluations(01:21:16) Nvidia AI Chips to Undergo Unusual U.S. Security Review Before Export to China(01:22:01) U.S. Authorities Shut Down Major China-Linked AI Tech Smuggling NetworkSynthetic Media & Art(01:24:01) RSL 1.0 has arrived, allowing publishers to ask AI companies pay to scrape content | The VergeSee Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

Permission To Speak Freely
Bonus Episode | Generalization Vs Generalization

Permission To Speak Freely

Play Episode Listen Later Dec 15, 2025 132:48


Aye, no long talk! Just check out this bonus episode. We love y'all and appreciate your continued support.   Music featured: “Gotta Do It” by Big K.R.I.T. Courtesy of Big K.R.I.T. / Multi Alumni / EMPIRE All rights belong to the original artist.

LessWrong Curated Podcast
“Weird Generalization & Inductive Backdoors” by Jorio Cocola, Owain_Evans, dylan_f

LessWrong Curated Podcast

Play Episode Listen Later Dec 14, 2025 17:32


Are they 18 yet?â„¢
Not seeing generalization? You might be the bottleneck.

Are they 18 yet?â„¢

Play Episode Listen Later Dec 10, 2025 26:19


If your therapy techniques only work when you're in the room, that's a problem.Many therapists unintentionally “gatekeep” their expertise and miss opportunities to boost carryover.It's the unexpected downside of being really good at direct clinical work. Don't get me wrong. Clinical judgment does matter. And some things can only be addressed by a trained clinician in a therapy room.But when every decision depends on your personal expertise and physical presence, you've made yourself the bottleneck.In this episode, I'll share how to make the shift towards clear, repeatable systems that others can follow. When you make your methods easier to teach, you make your work scalable, easier to delegate, and far more convincing to leadership.I'll tackle common misconceptions like:✅ “I can't delegate; I don't have direct reports.”✅ “I don't have time for consultation.”✅ “We never get enough time to work on skills.”Plus I share the three steps to making intervention “scalable” so your session plans can start doubling as consultation guides and training tools for others.In this episode, I mentioned Language Therapy Advance Foundations, my program that gives speech pathologists a framework for building language skills needed to thrive in school, social situations, and daily life. You can learn more about the program here: https://drkarenspeech.com/languagetherapyI also mentioned the School of Clinical Leadership, my program that helps related service providers develop scalable executive functioning strategies they can turn into schoolwide initiatives. You can learn more about the program here: https://drkarendudekbrannan.com/clinicalleadership We're thrilled to be sponsored by IXL. IXL's comprehensive teaching and learning platform for math, language arts, science, and social studies is accelerating achievement in 95 of the top 100 U.S. school districts. Loved by teachers and backed by independent research from Johns Hopkins University, IXL can help you do the following and more:Simplify and streamline technologySave teachers' timeReliably meet Tier 1 standardsImprove student performance on state assessments

K9 Detection Collaborative
Chocolate Cake, Green Eggs, and Ham: Analogies in Canine Detection & Generalization

K9 Detection Collaborative

Play Episode Listen Later Dec 2, 2025 45:09


What to listen for:Our hosts Robin Greubel and Stacy Barnett explore generalization as the foundation of reliable detection work.Together, they reveal generalization as extending far beyond simple obedience across locations. It encompasses odor variability, environmental context, behavioral chains, and handler presentation.They explain how explosive and narcotic sources vary like chocolate-cake recipes: different manufacturers, cutting agents, and absorption materials create distinct odor profiles.Dogs trained on limited sources may fail to recognize the "same" target odor prepared differently. That's why handlers must expose dogs to diverse training aids and seek out other teams' materials.Next, they talk behavioral generalization. Does "search" mean the same thing in a familiar training field versus a novel parking lot, rubble pile, or aircraft? Robin and Stacy stress that context cues (vehicles, wilderness, buildings) and environmental distractions require deliberate proofing so dogs maintain focus regardless of setting, weather, or ambient noise.Robin describes her area-search class methodology, which emphasizes that handlers can proof refind behaviors solo by generalizing the chain across handler positions. You could even do jumping jacks or lie turtle-like after falling into a hole.The goal is stimulus control, which means the cue triggers the behavior everywhere, every time.Our hosts warn against training disengagement by repeatedly working in overly distracting environments (woods full of "trail mix") without first building a clean chain in controlled settings like big-box stores with clean floors.If dogs routinely self-employ or search lackadaisically, handlers must assess whether hides exceed the dog's skill level, reinforcement is insufficient, or engagement was never properly conditioned.Their green-eggs-and-ham framework captures the essence of generalization: master the skill (row your boat), then generalize it everywhere (here, there, everywhere).Key Topics:Odor Generalization Across Sources and Absorption Materials (01:41)Training-Aid Availability and Pairing New Sources (04:56)Directionals and Platform Generalization (FEMA, Rubble Piles) (12:40)Training for Test vs. Application (Go-Outs, Distance, Body Language) (16:51)Area Search, Refind/TFR and Robin's New Class (18:50)Search Cue Stimulus Control in Blank and Novel Areas (20:45)Context Cues, Vehicles, and High-Likelihood Targets (23:38)Distraction Management and Clean Behavior Chains (37:45)Green Eggs and Ham: Progression Plans for Young Dogs (42:56) Resources:Study about the need for generalization in Explosives Detection DogsEpisodes with Dr. Lauryn DeGreeffRobin's Area Search/Wilderness Dog ClassStacy's course on Reading Pre-alert BehaviorMake sure to register for Stacy's upcoming term!We want to hear from you:Check out the K9 Detection Collaborative FB page and comment on the episode post!You can follow us for notifications of upcoming episodes, find us at k9detectioncollaborative.com

The MAD Podcast with Matt Turck
What's Next for AI? OpenAI's Łukasz Kaiser (Transformer Co-Author)

The MAD Podcast with Matt Turck

Play Episode Listen Later Nov 26, 2025 65:25


We're told that AI progress is slowing down, that pre-training has hit a wall, that scaling laws are running out of road. Yet we're releasing this episode in the middle of a wild couple of weeks that saw GPT-5.1, GPT-5.1 Codex Max, fresh reasoning modes and long-running agents ship from OpenAI — on top of a flood of new frontier models elsewhere. To make sense of what's actually happening at the edge of the field, I sat down with someone who has literally helped define both of the major AI paradigms of our time.Łukasz Kaiser is one of the co-authors of “Attention Is All You Need,” the paper that introduced the Transformer architecture behind modern LLMs, and is now a leading research scientist at OpenAI working on reasoning models like those behind GPT-5.1. In this conversation, he explains why AI progress still looks like a smooth exponential curve from inside the labs, why pre-training is very much alive even as reinforcement-learning-based reasoning models take over the spotlight, how chain-of-thought actually works under the hood, and what it really means to “train the thinking process” with RL on verifiable domains like math, code and science. We talk about the messy reality of low-hanging fruit in engineering and data, the economics of GPUs and distillation, interpretability work on circuits and sparsity, and why the best frontier models can still be stumped by a logic puzzle from his five-year-old's math book.We also go deep into Łukasz's personal journey — from logic and games in Poland and France, to Ray Kurzweil's team, Google Brain and the inside story of the Transformer, to joining OpenAI and helping drive the shift from chatbots to genuine reasoning engines. Along the way we cover GPT-4 → GPT-5 → GPT-5.1, post-training and tone, GPT-5.1 Codex Max and long-running coding agents with compaction, alternative architectures beyond Transformers, whether foundation models will “eat” most agents and applications, what the translation industry can teach us about trust and human-in-the-loop, and why he thinks generalization, multimodal reasoning and robots in the home are where some of the most interesting challenges still lie.OpenAIWebsite - https://openai.comX/Twitter - https://x.com/OpenAIŁukasz KaiserLinkedIn - https://www.linkedin.com/in/lukaszkaiser/X/Twitter - https://x.com/lukaszkaiserFIRSTMARKWebsite - https://firstmark.comX/Twitter - https://twitter.com/FirstMarkCapMatt Turck (Managing Director)Blog - https://mattturck.comLinkedIn - https://www.linkedin.com/in/turck/X/Twitter - https://twitter.com/mattturck(00:00) – Cold open and intro(01:29) – “AI slowdown” vs a wild week of new frontier models(08:03) – Low-hanging fruit: infra, RL training and better data(11:39) – What is a reasoning model, in plain language?(17:02) – Chain-of-thought and training the thinking process with RL(21:39) – Łukasz's path: from logic and France to Google and Kurzweil(24:20) – Inside the Transformer story and what “attention” really means(28:42) – From Google Brain to OpenAI: culture, scale and GPUs(32:49) – What's next for pre-training, GPUs and distillation(37:29) – Can we still understand these models? Circuits, sparsity and black boxes(39:42) – GPT-4 → GPT-5 → GPT-5.1: what actually changed(42:40) – Post-training, safety and teaching GPT-5.1 different tones(46:16) – How long should GPT-5.1 think? Reasoning tokens and jagged abilities(47:43) – The five-year-old's dot puzzle that still breaks frontier models(52:22) – Generalization, child-like learning and whether reasoning is enough(53:48) – Beyond Transformers: ARC, LeCun's ideas and multimodal bottlenecks(56:10) – GPT-5.1 Codex Max, long-running agents and compaction(1:00:06) – Will foundation models eat most apps? The translation analogy and trust(1:02:34) – What still needs to be solved, and where AI might go next

The Pepper & Dylan Show
November 24, 2025 - Wicked Review, Gender Generalizations, and Butter Questions

The Pepper & Dylan Show

Play Episode Listen Later Nov 24, 2025 25:02


Dylan's thoughts on the new Wicked: For Good. We discuss the tragedies that happened while filming the original Wizard Of Oz movie. We make some gender generalizations. Are woman able to see HD or do they just not care? Are woman less likely to talk to you while they're in the bathroom? Another major concert announce! The warning labels that scare us the most. Pepper has a question about butter. Can it be classified as just a spread?

Holmberg's Morning Sickness
11-21-25 - Holmberg Is Finding That Being His Own Booking Agent Is Hard - Hard Reality Of Car Repair Has Struck John Again As He Thought He Had A Win In Replacing His Wipers - John Makes A Generalization About Car Guys And Nails It

Holmberg's Morning Sickness

Play Episode Listen Later Nov 21, 2025 42:37


11-21-25 - Holmberg Is Finding That Being His Own Booking Agent Is Hard - Hard Reality Of Car Repair Has Struck John Again As He Thought He Had A Win In Replacing His Wipers - John Makes A Generalization About Car Guys And Nails ItSee Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

Holmberg's Morning Sickness - Arizona
11-21-25 - Holmberg Is Finding That Being His Own Booking Agent Is Hard - Hard Reality Of Car Repair Has Struck John Again As He Thought He Had A Win In Replacing His Wipers - John Makes A Generalization About Car Guys And Nails It

Holmberg's Morning Sickness - Arizona

Play Episode Listen Later Nov 21, 2025 42:37


11-21-25 - Holmberg Is Finding That Being His Own Booking Agent Is Hard - Hard Reality Of Car Repair Has Struck John Again As He Thought He Had A Win In Replacing His Wipers - John Makes A Generalization About Car Guys And Nails ItSee Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

Animal Training Academy
Generalization, Growth & Good Faith Learning – Ellen Yoakum [Episode 266]

Animal Training Academy

Play Episode Listen Later Nov 3, 2025 49:05


In this insightful conclusion to our two-part conversation, Ellen Yoakum—Certified Separation Anxiety Pro Behavior Consultant, KPA CTP, and Behavior Consultant with Pet Harmony—returns to explore how we can thoughtfully apply the learning principles we use with our animal learners to ourselves and the humans we work with. Building on the foundations of Part One, Ellen and Ryan dive into the complexities of generalization—how humans, much like our non-human learners, can struggle to transfer skills across contexts. From communication and empathy to client coaching and professional growth, Ellen offers compassionate strategies for building fluency, resilience, and sustainability in both behavior change and life. Together, they explore: ✅ Translating behavioral principles from dogs to humans—and ourselves ✅ Creating safe learning spaces for clients, colleagues, and trainers alike ✅ How understanding "pain points" can reshape client plans and improve outcomes ✅ Teaching for generalization without overwhelming learners ✅ Rethinking perfection and building skills for recovery when "life gets lifey" Ellen's reflections on generalization, empathy, and sustainable growth remind us that training isn't just about the animals—it's about the entire learning ecosystem. Her insights leave us inspired to meet our learners where they are, celebrate progress over perfection, and approach every interaction—human or non-human—with curiosity and care. Links Enrichment for the Real World Petharmonytraining.com Pet Harmony on Instagram and Facebook Pet Harmony Pro on Instagram and TikTok

By Kids, For Kids Story Time
The One Orc Conclusion - The Hasty Generalization logical fallacy

By Kids, For Kids Story Time

Play Episode Listen Later Nov 2, 2025 15:15


The One-Orc Conclusion: Judging a Village by One Bad Apple!

NeuroRadio
#99 Generalization

NeuroRadio

Play Episode Listen Later Oct 23, 2025 151:29


FSUの吉良信一郎さん(@ShinichiroKira)ゲスト回。UW Shadlen研への留学、サル電気生理からマウス(HMS Harvey研)へ、新たなラボでの研究の方向性など (10/5 収録)Show Notes (番組HP):Florida State University吉良ラボHPFlorida State University (FSU)で新たに研究室を立ち上げました。脳が過去の経験をもとに、新しい環境でどのように適切な意思決定を行うか (Generalization) にフォーカスして研究を進めていく予定です。熱意あるポスドク・大学院生を募集しています。ご興味のある方は、ぜひお気軽にご連絡ください!NRジャーナルクラブ回Lake ConferencesFlorida Panhandleタラハシー市シアトルのあるワシントン州の州都はオリンピア。タコマは軍基地がある。Florida State University フロリダ州立大学服部さん (Scripps, The University of Florida フロリダ大学)Seminoles (FSU) vs Gators (UF)奥山 輝大さん (東大 定量研)北沢 茂先生 (現・阪大)宇賀 貴紀先生(現・山梨大)三谷 明範さん(Komiyama研PhD->Google->-Artera>Apple)の宇賀研での仕事Saltzman/Newsome1992 最近の2光子Stim論文とかでこれを引いてないのはモグリだと思ってる(萩)Shadlen Newsome 1998 2001 一番刺さったShadlen et al 1996 (吉良)Goldman-Rakicのobituary in Nature. Funahashi/Goldman-Rakic1989が特によい(萩) Logothetis いまは中国にいるチュービンゲンのゴタゴタ Natureの記事、Scienceの記事Richard AxelDoris Tsaoのマウス、Tree shrewの仕事Davide ZoccolanJim DicarloMichael ShadlenPaul GlimcherRichard AndersenMichel GoldbergJacqueline GottliebBill NewsomeMante, Sussilo, Shenoy, NewsomeDaeyeol Lee このGameTheory/Social Decision-Makingの総説の印象が強い(萩)Xiao Jing WangそういえばLIPあんまり重要でなさそうというAlexHuk論文を議論してもよかったなと思いました。その後の展開をちゃんと追ってないので、LIP詳しい人のコメントを聞いてみたかったのでもしよかったらここで。(萩)Zhou & Freedman は抑制したLIPの応答視覚領域(Response Field)の中に視覚刺激(ランダムドット)がある場合とターゲットがある場合を比較して、前者の方が後者より正答率が顕著に低下することを示しました。LIP抑制ではないですが、Fetsch et al はMTが抑制されても残存領域の代替により、意思決定機能は速やかに回復することを示しました。同様の現象は齧歯類でも報告があります(Kawai et al, Mizes et al, Hong et al, Quintana et al)。私自身も実験を通じて経験しており、今後の研究テーマ(脳の機能回復メカニズム)として掲げています。(吉良)Eb Fetz (UW)Roozbeh Kiani (NYU)Tim Hanks (UC Davis)Anne Churchland (UCLA)Mehrdad Jazayeri (MIT)Josh Gold (U Penn)Tianming Yang (ION 中国科学院 )Probabilistic reasoning by neuronsNeural Implementation of Wald's Sequential Probability Ratioランダムドットタスク今年出したヒトのPsychophysics論文NIH RexPsychoPyリアルタイムOSMixed Selectivity 論文Tirin MooreFEF:Frontal eye field,眼球運動に対応する運動野だがselective attentionもやっているNHP NeuropixelSteppingとRampingの議論Neuropixel記録論文: Drift-Diffusion Modelの実証Elizabeth BaffaloMemory-Guided Saccade TaskStefano FusiRigotti et al (Mixed selectivity)Barnerdi et al (CCGP: Cross Condition Generalization Performance)Schemaを使ったCategorizationの簡潔な総説を五十嵐さんが書いているSaltzman & FusiCourellis et al (ヒトの神経活動をCCGPで解析)Attila Losonczy  (現在はUniversity of Texas South Western)Larry AbbottKen MillerLiam PaninskiAllen D3Laura DriscollNathaniel DawPeter DayanThalamusのU-19 GrantShaul Druckmann大久保 達夫さん NR出演回加藤 郁佳さん NR出演回Chris HarveyHarveyTank 2012 2012年3月なのでまさにSalk滞在中に出版されて読んでた(萩)HattoriさんのRSCの論文 (Hattori 2019)JeremiahのmPFCの論文 (Bari 2019)Default mode networkDelay (non) match to sample taskShadlenのOdorの論文JerryChenのS1/S2の論文, S1しか見てない論文Voitov & Mrsic-Flogel論文 ちなみに前述のLosonczy研にいてNYに残った知り合いというのはこのIvan(萩)

K9 Detection Collaborative
Talking Detection with Steve White Pt. 2

K9 Detection Collaborative

Play Episode Listen Later Oct 21, 2025 50:04


What to listen for:“Search and rescue work is the volunteer profession that you pay to do and is one of the most stressful things you can ever do, because somebody else's life could depend on what you're doing.”In part 2 of their conversation with Steve White, Robin Greubel and Stacy Barnett ask about the development of Hydrated Intensive Tracking (HIT), which evolved from experiments with scent-in-a-bottle methods.Steve's breakthrough came at a U.S. Police Canine Association seminar when handlers lacked marker training skills. By hybridizing traditional food-in-footstep methods with spray lines, Steve discovered that dogs crossing pavement with spray present kept their heads lower even after the spray evaporated. It's classical conditioning at work!Steve's training philosophy emphasizes creating calm, methodical working dogs rather than frantic high-energy animals. He seeks dogs with "conditioned emotional responses" of focused steadiness. He believes that clearheaded dogs perform better in difficult urban environments. This approach influenced his article training, where teaching dogs to find tiny objects like washers creates precision that makes finding larger targets effortless.Robin and Stacy zero in on the importance of generalization and stimulus control. Dogs absolutely distinguish training from operations, requiring extensive work in operational environments. Steve advocates for the "Green Eggs and Ham" principle. That is, can your dog perform here, there, everywhere? Handlers often mistake lack of stimulus control for lack of behavior knowledge.His current work with the United States Police Canine Association's Best Practices Working Group aims to preserve police canine programs by shifting focus toward the irreplaceable value of dogs' olfactory capabilities while promoting cooperation-based control methods over force-dependent approaches.Key Topics:Search Dogs vs. Examination Dogs (01:40)Evolution of Hydrated Intensive Tracking (12:09)Classical Conditioning and Surface Work (17:47)Generalization and Stimulus Control (26:48)Training for Operational Environments (36:37)Takeaways (45:23)Resources:You can find Steve White:Proactive K9 WebsiteProactive K9 Website FormsUSPCA YouTube Channel: Where you can find Steve's three-part series on odor/scent fundamentals, a 1000-hour eyes presentation where he talks about the eight indicators of dogs being on odor, and Robin's presentations about the recipe for building a great training session.We want to hear from you:Check out the K9 Detection Collaborative FB page and comment on the episode post!K9Sensus Detection Dog Trainer AcademyK9Sensus Foundation can be found on Facebook and Instagram. We have a Trainer's Group on Facebook!Scentsabilities Nosework is also on Facebook. Here is a Facebook group you should join!Crystal Wing (CB K9) can be found here!You can follow us for notifications of upcoming episodes, find us at k9detectioncollaborative.com

The How to ABA Podcast
Natural Doesn't Mean Unplanned: The Art of Intentional Play-Based Teaching

The How to ABA Podcast

Play Episode Listen Later Oct 21, 2025 22:36


When we say “natural,” it's easy to imagine free play and child-led sessions where the therapist simply follows along. But natural doesn't mean unplanned. Behind every playful moment is an opportunity for intentional teaching, and the best play-based sessions are those that blend spontaneity with strategy.In this conversation, we talk about how to plan for play in a way that still feels natural and fun. We compare it to hosting a birthday party—you wouldn't invite a group of kids and hope for the best without games or structure! The same goes for ABA sessions. Following a child's lead doesn't mean letting them bounce from toy to toy. It means embedding purposeful teaching moments within what they love most.We share practical ways to create this balance, from using assessments like the Early Start Denver Model to identify meaningful goals to organizing the environment so motivation and learning opportunities flow naturally. You'll hear how we think through contriving the environment, following motivation rather than chaos, and using every playful moment with purpose.What's Inside:How to combine natural play with intentional teachingUsing assessments to guide play-based goalsTips for setting up the environment to boost engagement and learningMentioned in This Episode:Episode 118: Generalization and Maintenance of Skills in ABAHowToABA.com/joinHow to ABA on YouTubeFind us on FacebookFollow us on Instagram 

Neoborn And Andia Human Show
Learn from History - Avoid Civil War

Neoborn And Andia Human Show

Play Episode Listen Later Oct 14, 2025 60:23


Neoborn Caveman reflects on history's lessons to prevent civil war, blending satire with calls for humanity's preservation amid rising divisions. NC recounts Balkan atrocities under the Ustashas, from Prebilovci school massacres to Jasenovac horrors, highlighting generalizations' dangers and media distortions that fuel enmity. He critiques the Cloward-Piven strategy's overload tactics in modern immigration and crises, Ottoman legacies in Yugoslavia's fractures, Sunni-Shia Muslim divides and how they affect Western countries through immigration, and First Nations' overlooked slave-owning past, urging unity over rage to counter globalist techno-feudal cages.Music guests: InoxidablesKey TakeawaysHistory reveals patterns to avoid repeating civil wars and massacres.Generalizations breed hatred and overlook good people in every group.Media portrays events to manipulate perceptions and sow division.Governments and extremists exploit crises for control and collapse.Unity across differences builds stronger communities than walls or fences.Religious and ethnic labels mask shared humanity and common enemies.Atrocities like in Jasenovac show evil's scale when unchecked by reason.Immigration without checks risks importing unresolved conflicts.Past empires' conquests echo in today's border and identity struggles.Preserving stories counters falsified narratives and eternal enmities.Techno-feudalism threatens sovereignty more than open dialogue.Pro-humanity choices prioritize wholeness over emotional rage.Sound Bites"Never generalize. Just because a cat scratched you, it doesn't mean cats are evil.""Let's learn more about history to understand our chances in the future so we can have a better choice in the present.""If you falsify history you will create an everlasting enemy.""Civil war must be avoided at all cost.""Sharing the same land should be enough. We should respect and enjoy our uniqueness and the difference in each other.""Do you want this next level communist, fascist, technologist dystopia or shall we learn from history and avoid sparing children messaging others?""You are amazing. You are special. You are one of a kind. And you are worthy.""We should be one people. Not one people in division but one people united for the betterment of ourselves and humanity."Gather for unfiltered rambles at patreon.com/theneoborncavemanshow—free join, chats, lives.Keywords: history, civil war, generalizations, Ustashas, Balkans, Yugoslavia, Ottoman Empire, Sunnis, Shias, immigration, globalists, humanity, unity, genocide, Jasenovac, Pavelić, media manipulation, Trump, Cloward-Piven, First NationsHumanity centered satirical takes on the world & news + music - with a marble mouthed host.Free speech marinated in comedy.Supporting Purple Rabbits. Hosted on Acast. See acast.com/privacy for more information.

CEO Pulse Podcast
From CPA to Profit: Financial Tips for Real Estate Investors w/ Amanda Webster & Rafael Cortez

CEO Pulse Podcast

Play Episode Listen Later Oct 4, 2025 46:57


In today's episode, Rafael Cortez is joined by Amanda Webster, the Chief Revenue Officer at Accruity, a leading financial advisory firm. With over two decades of experience, Amanda shares her insights on how real estate investors and professionals can optimize their financial strategies to save money, grow their revenue, and navigate the complex world of taxes.Amanda Webster is an expert in financial strategy for real estate professionals, including investors, construction companies, brokerages, and property management firms. As a military spouse and mother of five, she brings the same resilience, dedication, and strategic thinking to her work as she does to her family life.We cover:1️⃣ Why your CPA might not be the right fit for your real estate business (and how to find the right one)2️⃣ How to legally offset tax liabilities and keep more of your profits

Autism Outreach
#247: How To Gamify Therapy with Lindsay Watson

Autism Outreach

Play Episode Listen Later Sep 23, 2025 30:02


Lindsay Watson, PT, CEO, and Co-Founder of Augment Therapy, is on a mission to blend augmented reality (AR) and virtual care to transform therapy. Augment Therapy offers interactive AR rehabilitation exercises and fun wellness games designed to encourage movement and improve outcomes at home and in person. With their ARWell PRO app, therapists can use the software during sessions and give patients free access at home, all while tracking progress through a customized, gamified platform.While Augment Therapy is currently used primarily by OTs and PTs, Lindsay shares exciting plans to expand into speech therapy. We also discuss the benefits of telehealth when applied intentionally and how leveraging technology can enhance repetition, generalization, and engagement—turning a tool that's often seen as a negative into a powerful ally for therapy success.#autism #speechtherapy What's Inside:What is Augment Therapy?How can Augmented Reality impact therapy.Blending expertise and virtual care.Mentioned In This Episode:Augment Therapy Join the aba speech connection  ABA Speech: HomeThe BriefAll your family's pressing concerns and questions, answered in one place. Mike...Listen on: Apple Podcasts Spotify

Machine Learning Street Talk
Deep Learning is Not So Mysterious or Different - Prof. Andrew Gordon Wilson (NYU)

Machine Learning Street Talk

Play Episode Listen Later Sep 19, 2025 123:48


Professor Andrew Wilson from NYU explains why many common-sense ideas in artificial intelligence might be wrong. For decades, the rule of thumb in machine learning has been to fear complexity. The thinking goes: if your model has too many parameters (is "too complex") for the amount of data you have, it will "overfit" by essentially memorizing the data instead of learning the underlying patterns. This leads to poor performance on new, unseen data. This is known as the classic "bias-variance trade-off" i.e. a balancing act between a model that's too simple and one that's too complex.**SPONSOR MESSAGES**—Tufa AI Labs is an AI research lab based in Zurich. **They are hiring ML research engineers!** This is a once in a lifetime opportunity to work with one of the best labs in EuropeContact Benjamin Crouzier - https://tufalabs.ai/ —Take the Prolific human data survey - https://www.prolific.com/humandatasurvey?utm_source=mlst and be the first to see the results and benchmark their practices against the wider community!—cyber•Fund https://cyber.fund/?utm_source=mlst is a founder-led investment firm accelerating the cybernetic economyOct SF conference - https://dagihouse.com/?utm_source=mlst - Joscha Bach keynoting(!) + OAI, Anthropic, NVDA,++Hiring a SF VC Principal: https://talent.cyber.fund/companies/cyber-fund-2/jobs/57674170-ai-investment-principal#content?utm_source=mlstSubmit investment deck: https://cyber.fund/contact?utm_source=mlst— Description Continued:Professor Wilson challenges this fundamental belief (fearing complexity). He makes a few surprising points:**Bigger Can Be Better**: massive models don't just get more flexible; they also develop a stronger "simplicity bias". So, if your model is overfitting, the solution might paradoxically be to make it even bigger.**The "Bias-Variance Trade-off" is a Misnomer**: Wilson claims you don't actually have to trade one for the other. You can have a model that is incredibly expressive and flexible while also being strongly biased toward simple solutions. He points to the "double descent" phenomenon, where performance first gets worse as models get more complex, but then surprisingly starts getting better again.**Honest Beliefs and Bayesian Thinking**: His core philosophy is that we should build models that honestly represent our beliefs about the world. We believe the world is complex, so our models should be expressive. But we also believe in Occam's razor—that the simplest explanation is often the best. He champions Bayesian methods, which naturally balance these two ideas through a process called marginalization, which he describes as an automatic Occam's razor.TOC:[00:00:00] Introduction and Thesis[00:04:19] Challenging Conventional Wisdom[00:11:17] The Philosophy of a Scientist-Engineer[00:16:47] Expressiveness, Overfitting, and Bias[00:28:15] Understanding, Compression, and Kolmogorov Complexity[01:05:06] The Surprising Power of Generalization[01:13:21] The Elegance of Bayesian Inference[01:33:02] The Geometry of Learning[01:46:28] Practical Advice and The Future of AIProf. Andrew Gordon Wilson:https://x.com/andrewgwilshttps://cims.nyu.edu/~andrewgw/https://scholar.google.com/citations?user=twWX2LIAAAAJ&hl=en https://www.youtube.com/watch?v=Aja0kZeWRy4 https://www.youtube.com/watch?v=HEp4TOrkwV4 TRANSCRIPT:https://app.rescript.info/public/share/H4Io1Y7Rr54MM05FuZgAv4yphoukCfkqokyzSYJwCK8Hosts:Dr. Tim Scarfe / Dr. Keith Duggar (MIT Ph.D)REFS:Deep Learning is Not So Mysterious or Different [Andrew Gordon Wilson]https://arxiv.org/abs/2503.02113Bayesian Deep Learning and a Probabilistic Perspective of Generalization [Andrew Gordon Wilson, Pavel Izmailov]https://arxiv.org/abs/2002.08791Compute-Optimal LLMs Provably Generalize Better With Scale [Marc Finzi, Sanyam Kapoor, Diego Granziol, Anming Gu, Christopher De Sa, J. Zico Kolter, Andrew Gordon Wilson]https://arxiv.org/abs/2504.15208

The Misfit Behaviorists - Practical Strategies for Special Education and ABA Professionals
Ep. 65: How to Teach Maintenance and Generalization of Skills in ABA and Special Education

The Misfit Behaviorists - Practical Strategies for Special Education and ABA Professionals

Play Episode Listen Later Sep 16, 2025 14:44 Transcription Available


If it's not generalizing, it's not functional. In this episode, we break down what true mastery looks like, how to plan for maintenance from the start, and practical ways to “teach loosely” so skills stick across people, places, and materials.

Podcast Notes Playlist: Latest Episodes
Physics Absorbed Artificial Intelligence & (Maybe) Consciousness

Podcast Notes Playlist: Latest Episodes

Play Episode Listen Later Sep 9, 2025


Theories of Everything with Curt Jaimungal ✓ Claim : Read the notes at at podcastnotes.org. Don't forget to subscribe for free to our newsletter, the top 10 ideas of the week, every Monday --------- As a listener of TOE you can get a special 20% off discount to The Economist and all it has to offer! Visit https://www.economist.com/toe MIT physicist Max Tegmark argues AI now belongs inside physics—and that consciousness will be next. He separates intelligence (goal-achieving behavior) from consciousness (subjective experience), sketches falsifiable experiments using brain-reading tech and rigorous theories (e.g., IIT/φ), and shows how ideas like Hopfield energy landscapes make memory “feel” like physics. We get into mechanistic interpretability (sparse autoencoders), number representations that snap into clean geometry, why RLHF mostly aligns behavior (not goals), and the stakes as AI progress accelerates from “underhyped” to civilization-shaping. It's a masterclass on where mind, math, and machines collide. Join My New Substack (Personal Writings): https://curtjaimungal.substack.com Listen on Spotify: https://open.spotify.com/show/4gL14b92xAErofYQA7bU4e Timestamps: - 00:00 - Why AI is the New Frontier of Physics - 09:38 - Is Consciousness Just a Byproduct of Intelligence? - 16:43 - A Falsifiable Theory of Consciousness? (The MEG Helmet Experiment) - 27:34 - Beyond Neural Correlates: A New Paradigm for Scientific Inquiry - 38:40 - Humanity: The Masters of Underestimation (Fermi's AI Analogy) - 51:27 - What Are an AI's True Goals? (The Serial Killer Problem) - 1:03:42 - Fermat's Principle, Entropy, and the Physics of Goals - 1:15:52 - Eureka Moment: When an AI Discovered Geometry on Its Own - 1:30:01 - Refuting the "AI Doomers": We Have More Agency Than We Think Links mentioned: - Max's Papers: https://scholar.google.com/citations?user=eBXEZxgAAAAJ&hl=en - Language Models Use Trigonometry to Do Addition [Paper]: https://arxiv.org/abs/2502.00873 - Generalization from Starvation [Paper]: https://arxiv.org/abs/2410.08255 - Geoffrey Hinton [TOE]: https://youtu.be/b_DUft-BdIE - Michael Levin [TOE]: https://youtu.be/c8iFtaltX-s - Iceberg of Consciousness [TOE]: https://youtu.be/65yjqIDghEk - Improved Measures of Integrated Information [Paper]: https://arxiv.org/abs/1601.02626 - David Kaiser [TOE]: https://youtu.be/_yebLXsIdwo - Iain McGilchrist [TOE]: https://youtu.be/Q9sBKCd2HD0 - Elan Barenholtz & William Hahn [TOE]: https://youtu.be/A36OumnSrWY - Daniel Schmachtenberger [TOE]: https://youtu.be/g7WtcTATa2U - Ted Jacobson [TOE]: https://youtu.be/3mhctWlXyV8 - The “All Possible Paths” Myth [TOE]: https://youtu.be/XcY3ZtgYis0 SUPPORT: - Become a YouTube Member (Early Access Videos): https://www.youtube.com/channel/UCdWIQh9DGG6uhJk8eyIFl1w/join - Support me on Patreon: https://patreon.com/curtjaimungal - Support me on Crypto: https://commerce.coinbase.com/checkout/de803625-87d3-4300-ab6d-85d4258834a9 - Support me on PayPal: https://www.paypal.com/donate?hosted_button_id=XUBHNMFXUX5S4 SOCIALS: - Twitter: https://twitter.com/TOEwithCurt - Discord Invite: https://discord.com/invite/kBcnfNVwqs Guests do not pay to appear. Theories of Everything receives revenue solely from viewer donations, platform ads, and clearly labelled sponsors; no guest or associated entity has ever given compensation, directly or through intermediaries. #science Learn more about your ad choices. Visit megaphone.fm/adchoices

Podcast Notes Playlist: Nutrition
Physics Absorbed Artificial Intelligence & (Maybe) Consciousness

Podcast Notes Playlist: Nutrition

Play Episode Listen Later Sep 9, 2025 109:53


Theories of Everything with Curt Jaimungal ✓ Claim Key Takeaways  Conditions like depression, bipolar disorder, and schizophrenia may be driven in part by metabolic dysfunction in the brainNeuroinflammation is real, but fasting and a ketogenic diet can help The benefits of supplementing exogenous ketones:(1) Quick energy – they give your body a fast fuel source, especially for the brain and muscles(2)  Support ketosis – they can help raise blood ketone levels even if you're not fully on a strict keto dietBenefits of fasting: helps to augment the control of the immune system, relaxes the gut and enables the body's repair processes to occur, reduces the body's general state of inflammation There is a ketone-synergistic effect when delivering caffeine with MCT; it stimulates lipolysis and also fat oxidation in the liver The short-list of essential supplements:CoQ10, creatine, ketones, vitamin D, and melatoninThe benefits of metformin and GLP-1 drugs may arise from their influence on metabolic functionA low-carb Mediterranean-style diet is conducive to upgrading your metabolic machinery while keeping biomarkers in checkDiet: No sugar, no starch, fibrous vegetables, aim for 25% of carbohydrates consumed should be from fiber, high-protein + low glycemic breakfast and lunch, then a pound of protein for dinner with some fibrous vegetables The protocol and surprising benefits of ‘Sardine Fasting': Eat 1-2 cans of sardines per day for one week; can be repeated monthly or as needed. May need to supplement with vitamin C and magnesium  Why it helps: Provides essential nutrients and omega-3s while keeping calories/protein low enough to activate autophagy, support immunity, fight brain fog, and promote overall metabolic health Movement is critical for optimal metabolic health; get outside and walk first thing in the morning, and try to move after dinner for the sake of glucose metabolism Read the full notes @ podcastnotes.orgAs a listener of TOE you can get a special 20% off discount to The Economist and all it has to offer! Visit https://www.economist.com/toe MIT physicist Max Tegmark argues AI now belongs inside physics—and that consciousness will be next. He separates intelligence (goal-achieving behavior) from consciousness (subjective experience), sketches falsifiable experiments using brain-reading tech and rigorous theories (e.g., IIT/φ), and shows how ideas like Hopfield energy landscapes make memory “feel” like physics. We get into mechanistic interpretability (sparse autoencoders), number representations that snap into clean geometry, why RLHF mostly aligns behavior (not goals), and the stakes as AI progress accelerates from “underhyped” to civilization-shaping. It's a masterclass on where mind, math, and machines collide. Join My New Substack (Personal Writings): https://curtjaimungal.substack.com Listen on Spotify: https://open.spotify.com/show/4gL14b92xAErofYQA7bU4e Timestamps: - 00:00 - Why AI is the New Frontier of Physics - 09:38 - Is Consciousness Just a Byproduct of Intelligence? - 16:43 - A Falsifiable Theory of Consciousness? (The MEG Helmet Experiment) - 27:34 - Beyond Neural Correlates: A New Paradigm for Scientific Inquiry - 38:40 - Humanity: The Masters of Underestimation (Fermi's AI Analogy) - 51:27 - What Are an AI's True Goals? (The Serial Killer Problem) - 1:03:42 - Fermat's Principle, Entropy, and the Physics of Goals - 1:15:52 - Eureka Moment: When an AI Discovered Geometry on Its Own - 1:30:01 - Refuting the "AI Doomers": We Have More Agency Than We Think Links mentioned: - Max's Papers: https://scholar.google.com/citations?user=eBXEZxgAAAAJ&hl=en - Language Models Use Trigonometry to Do Addition [Paper]: https://arxiv.org/abs/2502.00873 - Generalization from Starvation [Paper]: https://arxiv.org/abs/2410.08255 - Geoffrey Hinton [TOE]: https://youtu.be/b_DUft-BdIE - Michael Levin [TOE]: https://youtu.be/c8iFtaltX-s - Iceberg of Consciousness [TOE]: https://youtu.be/65yjqIDghEk - Improved Measures of Integrated Information [Paper]: https://arxiv.org/abs/1601.02626 - David Kaiser [TOE]: https://youtu.be/_yebLXsIdwo - Iain McGilchrist [TOE]: https://youtu.be/Q9sBKCd2HD0 - Elan Barenholtz & William Hahn [TOE]: https://youtu.be/A36OumnSrWY - Daniel Schmachtenberger [TOE]: https://youtu.be/g7WtcTATa2U - Ted Jacobson [TOE]: https://youtu.be/3mhctWlXyV8 - The “All Possible Paths” Myth [TOE]: https://youtu.be/XcY3ZtgYis0 SUPPORT: - Become a YouTube Member (Early Access Videos): https://www.youtube.com/channel/UCdWIQh9DGG6uhJk8eyIFl1w/join - Support me on Patreon: https://patreon.com/curtjaimungal - Support me on Crypto: https://commerce.coinbase.com/checkout/de803625-87d3-4300-ab6d-85d4258834a9 - Support me on PayPal: https://www.paypal.com/donate?hosted_button_id=XUBHNMFXUX5S4 SOCIALS: - Twitter: https://twitter.com/TOEwithCurt - Discord Invite: https://discord.com/invite/kBcnfNVwqs Guests do not pay to appear. Theories of Everything receives revenue solely from viewer donations, platform ads, and clearly labelled sponsors; no guest or associated entity has ever given compensation, directly or through intermediaries. #science Learn more about your ad choices. Visit megaphone.fm/adchoices

Theories of Everything with Curt Jaimungal
Physics Absorbed Artificial Intelligence & (Maybe) Consciousness

Theories of Everything with Curt Jaimungal

Play Episode Listen Later Sep 3, 2025 109:53


As a listener of TOE you can get a special 20% off discount to The Economist and all it has to offer! Visit https://www.economist.com/toe MIT physicist Max Tegmark argues AI now belongs inside physics—and that consciousness will be next. He separates intelligence (goal-achieving behavior) from consciousness (subjective experience), sketches falsifiable experiments using brain-reading tech and rigorous theories (e.g., IIT/φ), and shows how ideas like Hopfield energy landscapes make memory “feel” like physics. We get into mechanistic interpretability (sparse autoencoders), number representations that snap into clean geometry, why RLHF mostly aligns behavior (not goals), and the stakes as AI progress accelerates from “underhyped” to civilization-shaping. It's a masterclass on where mind, math, and machines collide. Join My New Substack (Personal Writings): https://curtjaimungal.substack.com Listen on Spotify: https://open.spotify.com/show/4gL14b92xAErofYQA7bU4e Timestamps: - 00:00 - Why AI is the New Frontier of Physics - 09:38 - Is Consciousness Just a Byproduct of Intelligence? - 16:43 - A Falsifiable Theory of Consciousness? (The MEG Helmet Experiment) - 27:34 - Beyond Neural Correlates: A New Paradigm for Scientific Inquiry - 38:40 - Humanity: The Masters of Underestimation (Fermi's AI Analogy) - 51:27 - What Are an AI's True Goals? (The Serial Killer Problem) - 1:03:42 - Fermat's Principle, Entropy, and the Physics of Goals - 1:15:52 - Eureka Moment: When an AI Discovered Geometry on Its Own - 1:30:01 - Refuting the "AI Doomers": We Have More Agency Than We Think Links mentioned: - Max's Papers: https://scholar.google.com/citations?user=eBXEZxgAAAAJ&hl=en - Language Models Use Trigonometry to Do Addition [Paper]: https://arxiv.org/abs/2502.00873 - Generalization from Starvation [Paper]: https://arxiv.org/abs/2410.08255 - Geoffrey Hinton [TOE]: https://youtu.be/b_DUft-BdIE - Michael Levin [TOE]: https://youtu.be/c8iFtaltX-s - Iceberg of Consciousness [TOE]: https://youtu.be/65yjqIDghEk - Improved Measures of Integrated Information [Paper]: https://arxiv.org/abs/1601.02626 - David Kaiser [TOE]: https://youtu.be/_yebLXsIdwo - Iain McGilchrist [TOE]: https://youtu.be/Q9sBKCd2HD0 - Elan Barenholtz & William Hahn [TOE]: https://youtu.be/A36OumnSrWY - Daniel Schmachtenberger [TOE]: https://youtu.be/g7WtcTATa2U - Ted Jacobson [TOE]: https://youtu.be/3mhctWlXyV8 - The “All Possible Paths” Myth [TOE]: https://youtu.be/XcY3ZtgYis0 SUPPORT: - Become a YouTube Member (Early Access Videos): https://www.youtube.com/channel/UCdWIQh9DGG6uhJk8eyIFl1w/join - Support me on Patreon: https://patreon.com/curtjaimungal - Support me on Crypto: https://commerce.coinbase.com/checkout/de803625-87d3-4300-ab6d-85d4258834a9 - Support me on PayPal: https://www.paypal.com/donate?hosted_button_id=XUBHNMFXUX5S4 SOCIALS: - Twitter: https://twitter.com/TOEwithCurt - Discord Invite: https://discord.com/invite/kBcnfNVwqs Guests do not pay to appear. Theories of Everything receives revenue solely from viewer donations, platform ads, and clearly labelled sponsors; no guest or associated entity has ever given compensation, directly or through intermediaries. #science Learn more about your ad choices. Visit megaphone.fm/adchoices

I Love Recruiting
Three Mistakes to Avoid As An Entrepreneur

I Love Recruiting

Play Episode Listen Later Jul 24, 2025 22:03


In this episode, Adam Roach and Jess Webber discuss the three costly mistakes that entrepreneurs often make when launching and scaling their businesses. They emphasize the importance of proper pricing, the dangers of perfectionism, and the pitfalls of being a generalist in coaching. The discussion provides actionable insights for coaches and entrepreneurs to avoid these common traps and achieve greater success.Sign up for the 10K Challenge hereTakeawaysMany coaches undercharge due to imposter syndrome.Pricing should reflect the transformation provided to clients.Confidence in pricing can lead to better client relationships.Perfectionism can hinder action and delay launches.Successful coaches often launch quickly and iterate based on feedback.Generalists struggle to attract clients due to vague messaging.Specialization allows for premium pricing and clearer value propositions.Clarity in messaging opens doors to opportunities.Adding value is more effective than discounting prices.Taking action is essential for building confidence and success.Chapters00:00 Common Mistakes Coaches Make07:20 The Perfection Trap14:46 The Danger of Generalization

Mind Matters
Revisiting Girls and ADHD

Mind Matters

Play Episode Listen Later Jul 11, 2025 35:47


Generalizations about ADHD haven't done girls any favors. ADHD often manifests very differently in girls than in boys. What causes parents, educators, and even doctors, to view the symptoms of ADHD differently with girls? We know that masking, and even variations in environments, often cause symptoms to be missed. How is that happening? We're revisiting the subject and chat with Stephen Hinshaw, a Distinguished Professor of Psychology at UC Berkeley and Professor of Psychiatry and Behavioral Sciences at UC San Francisco. Find out more about building neurodiversity-affirming schools, and about the courses we'll be offering soon, here. Stephen P. Hinshaw is Distinguished Professor of Psychology at UC Berkeley and Professor of Psychiatry and Behavioral Sciences at UC San Francisco. His focus is on developmental psychopathology, child and adolescent mental health (particularly ADHD), and the use of clinical trials to understand underlying mechanisms. He also actively investigates mental illness stigmatization and attempts to reduce such stigma. Hinshaw has authored over 400 articles, chapters, and commentaries, plus 12 books. He has won numerous national and international research awards, including the James McKeen Cattell Fellow Award from the Association for Psychological Science, the Distinguished Scientific Contributions Award from the American Psychological Association, and the Sarnat International Prize in Mental Health from the National Academy of Medicine. He was inducted into the American Academy of Arts and Sciences in 2021. His extensive media coverage includes the New York Times, Washington Post, Wall Street Journal, Today Show, CBS Evening News, ABC World News Tonight, and many more. BACKGROUND READING Berkeley, research, website, Straight Talk about ADHD in Girls: How to Help Your Daughter Thrive The Neurodiversity Podcast is on Facebook, Instagram, BlueSky, and you're invited to join our Facebook Group.

Conquer Athlete Podcast
CAP 263 Building Mental Toughness in Athletes | 19 June 2025

Conquer Athlete Podcast

Play Episode Listen Later Jun 19, 2025 32:31


Summary In this episode of the Conquer Athlete Podcast, hosts Ryan Bucciantini and Jason Leydon discuss various themes surrounding fitness, mental resilience, and the importance of mindset in athletic performance. They share personal anecdotes from their experiences in the fitness community, including the origins of the D-BAP philosophy, which emphasizes mental toughness and resilience. The conversation also touches on the challenges athletes face, the significance of gradual exposure to difficult tasks, and the camaraderie built through shared fitness experiences. In this conversation, Ryan Bucciantini and Jason Leydon explore the intricacies of mental resilience, emphasizing the importance of discomfort, self-awareness, and the distinction between need and want in training. They discuss how mental toughness can be developed through intentional practices and the necessity of reflection in training. The dialogue also touches on the balance between focus and generalization in an athlete's life, the risk of burnout, and the significance of planning and prioritization in achieving long-term success.   Takeaways Mindset resilience is crucial for athletes. Quitting can lead to negative self-talk. Gradual exposure to challenges prevents burnout. The D-BAP philosophy encourages mental toughness. Community and camaraderie enhance the fitness experience. Training should be balanced to avoid overexertion. Personal anecdotes enrich the discussion on fitness. The importance of frequency in mental training. Eating habits can impact athletic performance. The evolution of fitness culture is fascinating. Mental training is akin to physical training; both require exposure to discomfort. Understanding the difference between need and want is crucial for mental resilience. Self-awareness is the foundation for building mental strength. Mental toughness can be compartmentalized but should be applied across life domains. Burnout often stems from a lack of self-knowledge and poor planning. Intentional planning can help prevent burnout and maintain balance. Reflection is essential to ensure that training intentions are met. Athletes must learn to stack their priorities to avoid neglecting other life areas. The journey to mental resilience is complex and requires ongoing effort. Creating a supportive environment can enhance mental training outcomes. Topics Building Mental Toughness in Athletes The D-BAP Philosophy Explained Mindset Resilience: Key to Athletic Success From Burnout to Brilliance: A Fitness Journey Sound Bites "Don't be a pussy. Dbap!" "We thought we had to eat paleo!" "Those training days were crazy!" "I can go back to those days!" "D-BAP, Conqueror Athlete D-BAP shirt!" "It's all about frequency and exposure." "Mental burnout is part poor planning." "Changing the way you think is complex." Chapters 00:00 Introduction to the Conquer Athlete Podcast 04:37 Mindset Resilience and Mental Toughness 14:06 The D-BAP Philosophy and Its Origins 15:15 Building Mental Resilience Through Discomfort 18:04 Understanding Need vs. Want in Mental Training 20:09 The Importance of Self-Awareness in Resilience 23:00 Translating Mental Toughness Across Life Domains 27:02 The Balance of Focus and Generalization in Training 29:51 Preventing Burnout Through Intentional Planning 33:00 Reflection and Intention in Mental Training

K9 Detection Collaborative
Detection Dog Science with Secret Service Chemist Dr. Katylynn Sloan Pt 2

K9 Detection Collaborative

Play Episode Listen Later Jun 17, 2025 45:14


What to listen for:Robin Greubel and Crystal Wing return with the brilliant Katylynn Sloan for part two of their conversation on the intersection between chemistry and K9 detection!“Train with as much variety as often as you can, in the weirdest of circumstances you can possibly get to, with as many groups and as many people as you can.”Katylynn dives into one of the most misunderstood topics in the detection world: training aids. She breaks down why the term “pseudo” is falling out of favor, replaced by “alternative training aids,” and classifies them into four types: dilution, absorption, mimic, and vigilance. Each has its pros and pitfalls. Mimics, for instance, rely on human interpretation of what's important for a dog to smell—sometimes right, sometimes not. And vigilance aids? They're about maintaining search behavior, not the odor itself.Katylynn also emphasizes the importance of language. As a member of standards boards like ASB and OSAC, she helps shape definitions so handlers, trainers, and scientists are all speaking the same language in courtrooms and classrooms alike. It's not just about what you know, but how clearly you can explain it. Her advice to aspiring canine scientists? Master problem-solving, communication, and the humility to say, “I don't know.”She also touches on the challenges of HRD training, the evolving definitions of “real” odor, and how even “duh” science needs to be written down. Her message to handlers? Train broadly. Generalization is key. Get variety in odors, people, places, and aids—because you never know what the real world will throw at your dog.Dogs are really good. But they're even better when guided by thoughtful, curious humans willing to adapt, collaborate, and learn. Katylynn's passion isn't just informative—it's contagious.Key Topics:Revising Standards and Defining "Real" Odor (0:59)Introduction to Alternative Training Aids (03:27)Mimics and Human Interpretation in Training Aids (07:30)Applying Dilution/Absorption Concepts to HRD (15:14)Public Comment Process and Impact on Standards (17:18)Skills and Traits for Aspiring K9 Scientists (22:53)Final Takeaways: Variety, Generalization, and Collaboration (39:52)Resources:The manipulation of odor availability of training aids used in detection canine trainingAAFS Academy Standards BoardLindsay Waldrop's LabWe want to hear from you:Check out the K9 Detection Collaborative FB page and comment on the episode post!K9Sensus Detection Dog Trainer AcademyK9Sensus Foundation can be found on Facebook and Instagram. We have a Trainer's Group on Facebook!Scentsabilities Nosework is also on Facebook. Here is a Facebook group you should join!Crystal Wing (CB K9) can be found here!You can follow us for notifications of upcoming episodes, find us at k9detectioncollaborative.com

MLOps.community
Everything Hard About Building AI Agents Today

MLOps.community

Play Episode Listen Later Jun 13, 2025 47:02


Willem Pienaar and Shreya Shankar discuss the challenge of evaluating agents in production where "ground truth" is ambiguous and subjective user feedback isn't enough to improve performance.The discussion breaks down the three "gulfs" of human-AI interaction—Specification, Generalization, and Comprehension—and their impact on agent success.Willem and Shreya cover the necessity of moving the human "out of the loop" for feedback, creating faster learning cycles through implicit signals rather than direct, manual review.The conversation details practical evaluation techniques, including analyzing task failures with heat maps and the trade-offs of using simulated environments for testing.Willem and Shreya address the reality of a "performance ceiling" for AI and the importance of categorizing problems your agent can, can learn to, or will likely never be able to solve.// BioShreya ShankarPhD student in data management for machine learning.Willem PienaarWillem Pienaar, CTO of Cleric, is a builder with a focus on LLM agents, MLOps, and open source tooling. He is the creator of Feast, an open source feature store, and contributed to the creation of both the feature store and MLOps categories. Before starting Cleric, Willem led the open source engineering team at Tecton and established the ML platform team at Gojek, where he built high scale ML systems for the Southeast Asian decacorn.// Related Linkshttps://www.google.com/about/careers/applications/?utm_campaign=profilepage&utm_medium=profilepage&utm_source=linkedin&src=Online/LinkedIn/linkedin_pagehttps://cleric.ai/~~~~~~~~ ✌️Connect With Us ✌️ ~~~~~~~Catch all episodes, blogs, newsletters, and more: https://go.mlops.community/TYExploreMLOps Swag/Merch: [https://shop.mlops.community/]Connect with Demetrios on LinkedIn: /dpbrinkmConnect with Shreya on LinkedIn: /shrshnkConnect with Willem on LinkedIn: /willempienaarTimestamps:[00:00] Trust Issues in AI Data[04:49] Cloud Clarity Meets Retrieval[09:37] Why Fast AI Is Hard[11:10] Fixing AI Communication Gaps[14:53] Smarter Feedback for Prompts[19:23] Creativity Through Data Exploration[23:46] Helping Engineers Solve Faster[26:03] The Three Gaps in AI[28:08] Alerts Without the Noise[33:22] Custom vs General AI[34:14] Sharpening Agent Skills[40:01] Catching Repeat Failures[43:38] Rise of Self-Healing Software[44:12] The Chaos of Monitoring AI

Celebrate Kids Podcast with Dr. Kathy
Navigating the New Landscape of Education: Specialization vs. Generalization

Celebrate Kids Podcast with Dr. Kathy

Play Episode Listen Later Jun 11, 2025 16:05 Transcription Available


In this episode, Wayne Stender explores the profound idea that a good life is more important than simply having a good job. He and Dr. Kathy define a good life as one filled with fulfillment, connection, joy, commitment, growth, service, and learning. Dr. Kathy encourages listeners to reflect on what a "good job" means to them, emphasizing aspects like providing for family, flexibility, and personal maturity. The conversation aims to deepen the understanding of how our careers can align with our overall life satisfaction and personal development. Additionally, Wayne highlights the importance of awakening different smarts in children through art, mentioning the work of Creating a Masterpiece, which offers structured art education to help kids thrive creatively and emotionally. Tune in for insights on prioritizing life fulfillment while navigating career choices.

This Week in Machine Learning & Artificial Intelligence (AI) Podcast
Grokking, Generalization Collapse, and the Dynamics of Training Deep Neural Networks with Charles Martin - #734

This Week in Machine Learning & Artificial Intelligence (AI) Podcast

Play Episode Listen Later Jun 5, 2025 85:21


Today, we're joined by Charles Martin, founder of Calculation Consulting, to discuss Weight Watcher, an open-source tool for analyzing and improving Deep Neural Networks (DNNs) based on principles from theoretical physics. We explore the foundations of the Heavy-Tailed Self-Regularization (HTSR) theory that underpins it, which combines random matrix theory and renormalization group ideas to uncover deep insights about model training dynamics. Charles walks us through WeightWatcher's ability to detect three distinct learning phases—underfitting, grokking, and generalization collapse—and how its signature “layer quality” metric reveals whether individual layers are underfit, overfit, or optimally tuned. Additionally, we dig into the complexities involved in fine-tuning models, the surprising correlation between model optimality and hallucination, the often-underestimated challenges of search relevance, and their implications for RAG. Finally, Charles shares his insights into real-world applications of generative AI and his lessons learned from working in the field. The complete show notes for this episode can be found at https://twimlai.com/go/734.

Dr. John Vervaeke
What AI Can Never Be | John Vervaeke

Dr. John Vervaeke

Play Episode Listen Later May 30, 2025 57:04


Can artificial intelligence truly become wise? In this landmark lecture, John Vervaeke explores the future of AI through a lens few dare to examine: the limits of intelligence itself. He unpacks the critical differences between intelligence, rationality, reasonableness, and wisdom—terms often used interchangeably in discussions around AGI. Drawing from decades of research in cognitive science and philosophy, John argues that while large language models like ChatGPT demonstrate forms of generalized intelligence, they fundamentally lack core elements of human cognition: embodiment, caring, and participatory knowing. By distinguishing between propositional, procedural, perspectival, and participatory knowing, he reveals why the current paradigm of AI is not equipped to generate consciousness, agency, or true understanding. This lecture also serves as a moral call to action: if we want wise machines, we must first become wiser ourselves. Connect with a community dedicated to self-discovery and purpose, and gain deeper insights by joining our Patreon. — 00:00 Introduction: AI, AGI, and the Nature of Intelligence 02:00 What is General Intelligence? 04:30 LLMs and the Illusion of Generalization 07:00 The Meta-Problems of Intelligence: Anticipation & Relevance Realization 09:00 Relevance Realization: The Hidden Engine of Intelligence 11:30 How We Filter Reality Through Relevance 14:00 The Limits of LLMs: Predicting Text vs. Anticipating Reality 17:00 Four Kinds of Knowing: Propositional, Procedural, Perspectival, Participatory 23:00 Embodiment, Consciousness, and Narrative Identity 27:00 The Role of Attention, Care, and Autopoiesis 31:00 Culture as Niche Construction 34:00 Why AI Can't Participate in Meaning 37:00 The Missing Dimensions in LLMs 40:00 Rationality vs. Reasonableness 43:00 Self-Deception, Bias, and the Need for Self-Correction 46:00 Caring About How You Care: The Core of Rationality 48:00 Wisdom: Aligning Multiple Selves and Temporal Scales 53:00 The Social Obligation to Cultivate Wisdom 55:00 Alter: Cultivating Wisdom in an AI Future — The Vervaeke Foundation is committed to advancing the scientific pursuit of wisdom and creating a significant impact on the world. Become a part of our mission: https://vervaekefoundation.org/ Join Awaken to Meaning to explore practices that enhance your virtues and foster deeper connections with reality and relationships: https://awakentomeaning.com/ — Ideas, People, and Works Mentioned in this Episode: Jeff Hinton Jordan Peterson Keith Stanovich Michael Levin Stroop Effect Bertrand Russell Plato (Republic, Symposium) Predictive Processing Relevance Realization Spearman (1926) DeepMind (DeepSeek) — Follow John Vervaeke: https://johnvervaeke.com/ https://twitter.com/vervaeke_john https://www.youtube.com/@johnvervaeke https://www.patreon.com/johnvervaeke — Thank you for watching!  

Network Marketing Breakthroughs with Rob Sperry
Why didn't network marketing explode after the pandemic?

Network Marketing Breakthroughs with Rob Sperry

Play Episode Listen Later May 26, 2025 4:48


I've been asked this question a lot lately…“Why didn't network marketing EXPLODE from the pandemic and recession, like so many expected? Why are we seeing declines instead of massive growth?”Let's get into it…There's no single reason — it's a combo of things. Yes, the pandemic created a surge for many companies in 2020–2021. But what people don't talk about is that growth was artificial.It was more of a sugar rush than it was sustainable momentum. People were home.They had more time.More stimulus checks.Ecommerce in general saw record-breaking numbers.But then what happened?Life returned to “normal,” attention spans shrank, wallets tightened, and expectations changed.Here's what I believe are the top reasons we've seen double-digit declines across many product-based companies:1. Artificial Pandemic BumpMany companies grew from a moment, not a movement. Growth wasn't built on deep belief or long-term behavior—it was built on convenience and timing.2. Lack of InnovationA lot of companies got lazy during the good times. They stopped innovating. Compensation plans didn't evolve. Products didn't stay ahead of the curve. Messaging got stale. And when consumer priorities shifted, companies weren't ready. This is a GENERALIZATION. I know many of you did evolve.3. We're in the “Options Era”Let's say someone used to buy 100 things. Now their budget forces them to cut back. 10–30% of those go first. Did your product make the cut?Uber. Airbnb. Affiliate. Creator economy. Ecom dropshipping. All legit options now. If we don't create real, unique value, we lose.Those options weren't avail in 2008.4. Poor Customer RetentionMost companies still don't do a great job reselling the customer after the sale. The fortune isn't just in the follow-up—it's in the retention experience.5. False ExpectationsPeople heard “work from home” and thought “easy income.” They weren't ready for rejection, consistency, and leadership development. So they quit early.Lazy on Leadership DevelopmentWe stopped developing leaders because it was so easy!Companies and leaders have been back in the trenches. Some for the last 2 years but you don't just plant a seed and expect a tree to grow the next day. It takes time to see the fruits of your labor. We are starting to see some of those results already and we will continue to see more. This is the WEEDING out season.Listen up—go to www.rankandbanksystem.com right now. Why? Because if your network marketing check isn't where you want it, this is the fix. It's my 30-day Rank & Bank System—daily trainings from me, daily challenges that actually move the needle, and accountability that keeps you in the game. Here's the kicker: it's got a 100% money-back guarantee. You follow the system, show up for 30 days, and if your check doesn't grow, you pay nothing. Zero risk. Most programs drown you in fluff—this one's built to get you results fast, without the overwhelm. I've taken people from $400 a month to millions (not normal but possible) with this exact framework. You've got nothing to lose and everything to gain—go to www.rankandbanksystem.com and lock it in before you miss out. Let's make your bank account match your hustle.

Gun Dog It Yourself
Recall - Ep. 282. The "Unforced Force Fetch"

Gun Dog It Yourself

Play Episode Listen Later Apr 24, 2025 106:15


WATCH --> https://2ly.link/1zQfw In this episode, we explore a modern and thoughtful take on force fetch with Chris Arminini of Full Send Canine. Learn how free shaping, marker training, and the NePoPo system are changing the game for retriever training. Whether you're a seasoned handler or starting your first dog, this episode brings clarity and innovation to one of the most debated training topics out there.

We're Not Marketers
Winning at Win-Loss: Why real research beats sales rep excuses w/ Ryan Sorley

We're Not Marketers

Play Episode Listen Later Apr 3, 2025 40:36


Is your win-loss data just expensive fiction written by your sales team? In this episode, Ryan Sorley (ex-Forrester, ex-Gartner) joins the crew to expose the hard truth about buyer research. From ripped jeans costing deals to bootstrapping a business while juggling kids and payroll, Ryan doesn't hold back on his journey from corporate misery to specialized success. Oh, and we uncovered the stupidly simple reason most product marketers fail at research. Tune in for honest laughs, real insights, and maybe a wake-up call about your "data."In this episode, we're covering:How to run a solo business when you have kids and bills to payThe "Taylor Swift Squad" theory of business growthWhy most buyer research is just wishful thinking with fancy graphsThe Powder Blue Taurus Moment: when Ryan knew corporate life was BSWhy writing a business plan is a complete waste of time for solopreneursHow Ryan went from 0 to 100 clients with ZERO salespeopleCheck out his new book Blindspots on Amazon too! Timestamps:01:00 Introducing Ryan Sorley, Win-Loss Research Expert02:36 Are Product Marketers Actually Marketers?04:15 Ryan's Take on Product Marketing as Research06:45 The Importance of Intentional Research10:00 The "Superpower" of Win-Loss Analysis12:34 The HubSpot Ripped Jeans Story17:15 Ryan's Entrepreneurial Journey19:00 The NJ Turnpike Moment: When Ryan Knew Corporate Life Wasn't for Him21:00 How Ryan Discovered the Win-Loss Opportunity23:45 Ryan's Transition from Gartner to Entrepreneurship27:00 What's the Minimum Viable Plan for Going Solo?30:00 The "Squad Life" Approach to Client Relationships34:00 The Dark Times: Managing Cash Flow and Contractors39:00 Specialization vs. Generalization in Consulting42:45 Ryan's New Book: "Blind Spots" (Launch April 1st)44:30 Closing Remarks and FarewellShow Notes:Ryan's LinkedInBlindspots on AmazonHosted by Ausha. See ausha.co/privacy-policy for more information.

The Behavioral Observations Podcast with Matt Cicoria
Response to Name Interventions for Staff and Caregivers: Inside JABA 22

The Behavioral Observations Podcast with Matt Cicoria

Play Episode Listen Later Mar 31, 2025 87:46


This is not a show about teaching eye contact. We'll get to that in a bit. First though, I should note that the 22nd installment of the Inside JABA Series is coming out comically late. I apologize for getting us off schedule. The good news is that we already have a great paper to discuss for the 23rd Inside JABA episode that I think you're going to love, so I hope to have that one out later on in the spring. Back to this episode. Drs. Danny Conine and Jenn Fritz join me to discuss a paper Danny wrote with his colleagues called, "Evaluating a screening-to-intervention model with caregiver training for response to name among children with autism." There are so many great things about this paper, and listeners will be able to tell this from my enthusiasm in discussing it with Danny and Jenn. As I noted above, this is not about teaching eye contact, but rather, a more generalized repertoire of responding to one's name (RTN). We get into why these two things are different, and, as Danny tells it, RTN repertoires have many benefits that directly impact learning and safety. In this paper, he describes an elegant assessment and intervention that his research team implemented to develop RTN in the study's participants. In carrying out this study, they also employed a simple and effective assent withdrawal component, which we get into. Then, they took what the skills they developed in a clinic setting, and taught the participant's caregivers to implement RTN procedures at home. As such, this paper provides a great example of how to generalize skills across settings. Very cool! Along the way, Danny provides practical tips clinicians can consider for their own practice. All of this to say, I'm hoping you'll agree that the wait for this episode will be worth it! Resources discussed in this podcast: Conine, et al. (2025). Evaluating a screening-to-intervention model with caregiver training for response to name among children with autism. Conine, et al. (2020). Assessment and treatment of response to name for children with autism spectrum disorder: Toward an efficient intervention model. Conine, Vollmer, and Bolívar (2019). Response to name in children with autism: Treatment, generalization, and maintenance. BOP Session 212 with Tim Hackenberg. Luczynski and Hanley (2013). Prevention of problem behavior by teaching functional communication and self-control skills to preschoolers. The Verbal Behavior Approach, by Dr. Mary Barbera. Links to Danny's faculty page, Research Gate profile, LinkedIn, and his lab's Instagram. Jenn's faculty page, Research Gate profile, LinkedIn, and the UHCL ABA Program page. If you enjoy this episode, please consider sharing with friends and colleagues!

Machine Learning Street Talk
Sakana AI - Chris Lu, Robert Tjarko Lange, Cong Lu

Machine Learning Street Talk

Play Episode Listen Later Mar 1, 2025 97:54


We speak with Sakana AI, who are building nature-inspired methods that could fundamentally transform how we develop AI systems.The guests include Chris Lu, a researcher who recently completed his DPhil at Oxford University under Prof. Jakob Foerster's supervision, where he focused on meta-learning and multi-agent systems. Chris is the first author of the DiscoPOP paper, which demonstrates how language models can discover and design better training algorithms. Also joining is Robert Tjarko Lange, a founding member of Sakana AI who specializes in evolutionary algorithms and large language models. Robert leads research at the intersection of evolutionary computation and foundation models, and is completing his PhD at TU Berlin on evolutionary meta-learning. The discussion also features Cong Lu, currently a Research Scientist at Google DeepMind's Open-Endedness team, who previously helped develop The AI Scientist and Intelligent Go-Explore.SPONSOR MESSAGES:***CentML offers competitive pricing for GenAI model deployment, with flexible options to suit a wide range of models, from small to large-scale deployments. Check out their super fast DeepSeek R1 hosting!https://centml.ai/pricing/Tufa AI Labs is a brand new research lab in Zurich started by Benjamin Crouzier focussed on o-series style reasoning and AGI. They are hiring a Chief Engineer and ML engineers. Events in Zurich. Goto https://tufalabs.ai/**** DiscoPOP - A framework where language models discover their own optimization algorithms* EvoLLM - Using language models as evolution strategies for optimizationThe AI Scientist - A fully automated system that conducts scientific research end-to-end* Neural Attention Memory Models (NAMMs) - Evolved memory systems that make transformers both faster and more accurateTRANSCRIPT + REFS:https://www.dropbox.com/scl/fi/gflcyvnujp8cl7zlv3v9d/Sakana.pdf?rlkey=woaoo82943170jd4yyi2he71c&dl=0Robert Tjarko Langehttps://roberttlange.com/Chris Luhttps://chrislu.page/Cong Luhttps://www.conglu.co.uk/Sakanahttps://sakana.ai/blog/TOC:1. LLMs for Algorithm Generation and Optimization [00:00:00] 1.1 LLMs generating algorithms for training other LLMs [00:04:00] 1.2 Evolutionary black-box optim using neural network loss parameterization [00:11:50] 1.3 DiscoPOP: Non-convex loss function for noisy data [00:20:45] 1.4 External entropy Injection for preventing Model collapse [00:26:25] 1.5 LLMs for black-box optimization using abstract numerical sequences2. Model Learning and Generalization [00:31:05] 2.1 Fine-tuning on teacher algorithm trajectories [00:31:30] 2.2 Transformers learning gradient descent [00:33:00] 2.3 LLM tokenization biases towards specific numbers [00:34:50] 2.4 LLMs as evolution strategies for black box optimization [00:38:05] 2.5 DiscoPOP: LLMs discovering novel optimization algorithms3. AI Agents and System Architectures [00:51:30] 3.1 ARC challenge: Induction vs. transformer approaches [00:54:35] 3.2 LangChain / modular agent components [00:57:50] 3.3 Debate improves LLM truthfulness [01:00:55] 3.4 Time limits controlling AI agent systems [01:03:00] 3.5 Gemini: Million-token context enables flatter hierarchies [01:04:05] 3.6 Agents follow own interest gradients [01:09:50] 3.7 Go-Explore algorithm: archive-based exploration [01:11:05] 3.8 Foundation models for interesting state discovery [01:13:00] 3.9 LLMs leverage prior game knowledge4. AI for Scientific Discovery and Human Alignment [01:17:45] 4.1 Encoding Alignment & Aesthetics via Reward Functions [01:20:00] 4.2 AI Scientist: Automated Open-Ended Scientific Discovery [01:24:15] 4.3 DiscoPOP: LLM for Preference Optimization Algorithms [01:28:30] 4.4 Balancing AI Knowledge with Human Understanding [01:33:55] 4.5 AI-Driven Conferences and Paper Review

The Multifamily Wealth Podcast
#268: Why Specialization (Not Generalization) Removes Ambiguity Around Your Business... And Is The Key To Scaling Your Portfolio

The Multifamily Wealth Podcast

Play Episode Listen Later Feb 4, 2025 16:29


In this episode, I share why developing specialized skills, rather than becoming a generalist,  allows you to scale and grow your multifamily portfolio more effectively. I discuss why determining whether you are going to become someone who's excellent at finding deals, raising debt/equity, or operating deals is key to removing the ambiguity around your business.This makes it easier for others to invest in your deals, connect you with people who can bring value to your business, and potentially partner with you, as well.I also use an analogy where I talk about Stringer Bell and The Wire... so this will be a fun episode for all of you out there who are fans of the show, like myself. Are you looking to invest in real estate, but don't want to deal with the hassle of finding great deals, signing on debt, and managing tenants? Aligned Real Estate Partners provides investment opportunities to passive investors looking for the returns, stability, and tax benefits multifamily real estate offers, but without the work - join our investor club to be notified of future investment opportunities.Connect with Axel:Follow him on InstagramConnect with him on LinkedInSubscribe to our YouTube channelLearn more about Aligned Real Estate Partners

scaling wire removes ambiguity specialization generalization stringer bell aligned real estate partners
K9 Detection Collaborative
A Conversation with Denise Fenzi Part 1

K9 Detection Collaborative

Play Episode Listen Later Jan 14, 2025 46:03


What to listen for:“If I cannot, in five repetitions, isolate a variable down and get it where I want it, I do need to stop. That doesn't mean the next time we'll be successful. What that tells me is, don't do the same thing again and again and again. And there's been a lot of variation. Isolate that variable. It looks good now. Put it back in a chain, it doesn't look good. What have I done wrong?”In part one of this two-part discussion, Robin Greubel and Crystal Wing sit down with special guest Denise Fenzi, a trainer and educator who specializes in building cooperation, joy, and extreme precision in competition dog sports teams.“Patterns lead to confidence. Unpredictable leads to listening.”Famous for breaking down complex concepts into bite-sized lessons, Denise does a deep dive into a training philosophy that stresses adaptability and prioritizing the dog's wellbeing. She discusses how to build your dog's vocabulary via strategic repetition, as well as her approach to encouraging or disrupting behavioral patterns during training.“The more ways you generalize the behavior, the better the dog gets at those behaviors.”Denise shares why tools like choke chains and e-collars might be a thing of the past, and why tailored, humane options like aversives better enhance understanding and cooperation between you and your dog–which in turn fosters a bond built on trust and respect rather than pain and fear.“A behavior chain is simply a bunch of behaviors strung together, so you have to get comfortable with degrees of messiness.”Finally, Denise unravels the complexities of behavior chains and the balance between instinctual behaviors and structured learning. She walks us through the art of rewarding gradual progress and adapting training techniques to maintain a dog's confidence.Key Topics:Denise's Training Methodology (06:51)Applying Micro Behaviors To Full Training Scenarios (20:22)Generalizing Behaviors (26:13)Drive Versus Arousal (35:50)Resources:Fenzi Dog Sports AcademyWe want to hear from you:Check out the K9 Detection Collaborative FB page and comment on the episode post!K9Sensus Detection Dog Trainer AcademyK9Sensus Foundation can be found on Facebook and Instagram. We have a Trainer's Group on Facebook!Scentsabilities Nosework is also on Facebook. Here is a Facebook group you should join!Crystal Wing K9 Coach can be found here at CB K9 and here at Evolution Working Dog Club. Also, check out her Functional Obedience Class here.You can follow us for notifications of upcoming episodes, find us at k9detectioncollaborative.comJingle by: www.mavericksings.comAudio editing & other podcast services by: www.thepodcastman.com

The Hake Report
Christmas Adam with Bigg Bump | Mon 12-23-24

The Hake Report

Play Episode Listen Later Dec 23, 2024 114:51


Sanctuary cities, music, Christmas movies/songs, Ozone hole "healing," stereotypes. Isaac Woodard was drunk! Blacks involved in Emmett Till kidnapping!The Hake Report, Monday, December 23, 2024 AD Bigg Bump https://www.youtube.com/@biggbump | https://x.com/bigg_bump | https://www.instagram.com/bigg_bump // TIMESTAMPS * (0:00:00) Start* (0:01:28) Topics with Bigg Bump* (0:05:44) Hey, guys!* (0:07:24) ALEX, CA: FE debunked? NASA lies.* (0:13:00) ALEX: Drone problems* (0:14:26) Bigg Bump: Air Force base, Chinese spy?* (0:16:27) York City Council, PA… Gaddafi; Female City Council* (0:34:26) DANIEL, TX: Bigg Bump music; natural vs taught talent* (0:40:02) DANIEL: Eminem…* (0:41:46) TERRI, OR: Fave Xmas movie, song? The Ref. Cristina, Things Fall Apart* (0:50:07) Ozone hole healing… science* (1:00:35) Trump court rulings* (1:05:49) Patience…* (1:06:40) MAZE, OH: NAACP, Regulations, Lawsuits* (1:13:50) MAZE: Do you listen to your husband? Debt Ceiling* (1:17:47) RIGO, TX: "Don't judge a book by its cover"; Generalizations* (1:24:40) Sgt Isaac Woodard story* (1:32:18) Being black…* (1:40:33) Emmett Till* (1:50:55) Jose Feliciano - "Feliz Navidad" - 1970LINKSBLOG https://www.thehakereport.com/blog/2024/12/23/christmas-with-bigg-bump-mon-12-23-24PODCAST / Substack HAKE NEWS from JLP https://www.thehakereport.com/jlp-news/2024/12/23/hake-news-mon-12-23-24Hake is live M-F 9-11a PT (11-1CT/12-2ET) Call-in 1-888-775-3773 https://www.thehakereport.com/showVIDEO YouTube - Rumble* - Facebook - X - BitChute - Odysee*PODCAST Substack - Apple - Spotify - Castbox - Podcast Addict*SUPER CHAT on platforms* above or BuyMeACoffee, etc.SHOP - Printify (new!) - Spring (old!) - Cameo | All My LinksJLP Network:JLP - Church - TFS - Nick - Joel - Punchie Get full access to HAKE at thehakereport.substack.com/subscribe

In Sanity: A piece of mind
Episode 199 - The Jackal: A Life Alienating System

In Sanity: A piece of mind

Play Episode Listen Later Nov 18, 2024 18:37


In Nonviolent Communication (NVC), the "Jackal" represents a communication style that is often aggressive and judgmental and tends to reflect a mindset focused on blame, criticism, or demands. Here are some key characteristics of the Jackal mode of communication: Judgment: Jackal communication often involves labeling people, actions, or thoughts as good or bad, right or wrong. This creates a sense of separation and defensiveness. Blame: The Jackal tends to assign responsibility to others for feelings or situations, often leading to conflict and resentment. It focuses on finding fault rather than understanding. Criticism: Jackal communication typically involves expressing disapproval or dissatisfaction in a hurtful or dismissive way, which can discourage open dialogue. Demands: Jackal communication often involves making demands rather than requests, creating feelings of fear or obligation rather than fostering a collaborative atmosphere. Generalizations : Making broad statements about people's actions or character can oversimplify complex situations and lead to misunderstandings. Emotional Disconnection: Jackal communication often lacks empathy, leading to emotional distance and a lack of connection with others' feelings and needs. In contrast, the "Giraffe" symbolizes a more compassionate form of communication in NVC. It focuses on expressing feelings and needs while fostering understanding and connection. NVC aims to shift from Jackal to Giraffe communication for healthier and more constructive interactions.

“HR Heretics” | How CPOs, CHROs, Founders, and Boards Build High Performing Companies
Beyond Emily in Paris: Navigating French Business Culture and HR Politics with Ciara Lakhani

“HR Heretics” | How CPOs, CHROs, Founders, and Boards Build High Performing Companies

Play Episode Listen Later Oct 31, 2024 33:39


Former Dashlane HR leader Ciara Lakhani shares her unfiltered journey from New York to Paris tech, debunking myths about French business culture while revealing the art of cross-cultural leadership. From smoke breaks with works councils to earning trust through authenticity, Ciara offers practical wisdom on navigating international business dynamics and building bridges between American and French work cultures. *Email us your questions or topics for Kelli & Nolan: hrheretics@turpentine.co For coaching and advising inquire at https://kellidragovich.com/ HR Heretics is a podcast from Turpentine.

RX'D RADIO
E537: World Views and Model Building with Dr. Pat Davidson PhD

RX'D RADIO

Play Episode Listen Later Jul 29, 2024 80:43


Shallow welcomes Dr. Patrick Davidson as they discuss the role of academia in the fitness industry. They reflect on the cycles and trends in the industry and the need for patience and discernment.  Dr. Davidson shares his thoughts on specialization versus generalization and the need for a holistic approach. He also discusses the importance of establishing a worldview and how it informs his coaching and teaching.   https://www.drpatdavidson.net/ https://www.instagram.com/dr.patdavidson/?hl=en   Join the Pre-Script® Level 1 Opt-In list now. Learn more at https://www.pre-script.com/psl1   We've got a new sponsor!   Marek Health is a health optimization company that offers advanced blood testing, health coaching, and expert medical oversight. Our services can help you enhance your lifestyle, nutrition, and supplementation to medical treatment and care. https://marekhealth.com/rxd Code RXD   Don't miss the release of our newest educational community - The Pre-Script ® Collective! Join the community today at www.pre-script.com.   For other strength training, health, and injury prevention resources, check out our website, YouTube channel, and Instagram.   For more episodes, subscribe and tune in to our podcast. Also, make sure to sign up to our mailing list at www.pre-script.com to get the first updates on new programming releases.   You can also follow Dr. Jordan Shallow and Dr. Jordan Jiunta on Instagram!   Dr. Jordan Shallow: https://www.instagram.com/the_muscle_doc/ Dr. Jordan Jiunta: https://www.instagram.com/redwiteandjordan/   Role of Academia in the Fitness Industry (00:03:07) Influence of Science and Well-Produced Content (00:07:23) Specialization vs Generalization in the Fitness Industry (00:11:02) Importance of Establishing a Worldview (00:17:27) Communication and Error in Driving Progress (00:27:07) Acting on Ideas and Spite (00:31:44) Pushing Boundaries in Training (00:36:12) Reconciling the Experiential Brain with the Theoretical Brain (00:40:43) Embracing Physical Discomfort and Physiological Changes (00:42:48) Using the Partner Fit App for Measuring Training Intensity (00:45:31)

The Domonique Foxworth Show
Examining the Best Way to Develop Athletes With David Epstein

The Domonique Foxworth Show

Play Episode Listen Later Jul 1, 2024 27:15


David Epstein joins the show to talk Olympics and how athletes are developed around the world. Topics include generalization vs. specialization in sports, developing athletes in the U.S. vs. other countries, why athletes can play for longer than ever before, and more! 0:00 Welcome back to the Domonique Foxworth Show 1:13 Generalization vs. specialization in sports 2:54 What should the U.S. copy from other countries in sports? 11:08 Could NFL players play in the NBA and vice versa? 15:18 Why do athletes have longer careers now? Learn more about your ad choices. Visit megaphone.fm/adchoices

Lex Fridman Podcast
#426 – Edward Gibson: Human Language, Psycholinguistics, Syntax, Grammar & LLMs

Lex Fridman Podcast

Play Episode Listen Later Apr 17, 2024 180:32


Edward Gibson is a psycholinguistics professor at MIT and heads the MIT Language Lab. Please support this podcast by checking out our sponsors: - Yahoo Finance: https://yahoofinance.com - Listening: https://listening.com/lex and use code LEX to get one month free - Policygenius: https://policygenius.com/lex - Shopify: https://shopify.com/lex to get $1 per month trial - Eight Sleep: https://eightsleep.com/lex to get special savings Transcript: https://lexfridman.com/edward-gibson-transcript EPISODE LINKS: Edward's X: https://x.com/LanguageMIT TedLab: https://tedlab.mit.edu/ Edward's Google Scholar: https://scholar.google.com/citations?user=4FsWE64AAAAJ TedLab's YouTube: https://youtube.com/@Tedlab-MIT PODCAST INFO: Podcast website: https://lexfridman.com/podcast Apple Podcasts: https://apple.co/2lwqZIr Spotify: https://spoti.fi/2nEwCF8 RSS: https://lexfridman.com/feed/podcast/ YouTube Full Episodes: https://youtube.com/lexfridman YouTube Clips: https://youtube.com/lexclips SUPPORT & CONNECT: - Check out the sponsors above, it's the best way to support this podcast - Support on Patreon: https://www.patreon.com/lexfridman - Twitter: https://twitter.com/lexfridman - Instagram: https://www.instagram.com/lexfridman - LinkedIn: https://www.linkedin.com/in/lexfridman - Facebook: https://www.facebook.com/lexfridman - Medium: https://medium.com/@lexfridman OUTLINE: Here's the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time. (00:00) - Introduction (10:53) - Human language (14:59) - Generalizations in language (20:46) - Dependency grammar (30:45) - Morphology (39:20) - Evolution of languages (42:40) - Noam Chomsky (1:26:46) - Thinking and language (1:40:16) - LLMs (1:53:14) - Center embedding (2:19:42) - Learning a new language (2:23:34) - Nature vs nurture (2:30:10) - Culture and language (2:44:38) - Universal language (2:49:01) - Language translation (2:52:16) - Animal communication