POPULARITY
Hey folks, Alex here, let me catch you up! Most important news about this week came today, mid-show, OpenAI dropped GPT 5.4 Thinking (and 5.4 Pro), their latest flagship general model, less autistic than Codex 5.3, with 1M context, /fast mode and the ability to steet it mid-reasoning. We tested it live on the show, it's really a beast. Also, since last week, Anthropic said no to Department of War's ultimatum and it looks like they are being designated as supply chain risk, OpenAI swooped in to sign a deal with DoW and the internet went ballistic (Dario also had some .. choice words in a leaked memo!) On the Open Source front, the internet lost it's damn mind when a friend of the pod Junyang Lin, announced his departure from Qwen in a tweet, causing an uproar, and the CEO of Alibaba to intervene. Wolfram presented our new in-house wolfbench.ai and a lot more! P.S - We acknowledge the war in Iran, and wish a quick resolution, the safety of civilians on both sides. Yam had to run to the shelter multiple times during the show. ThursdAI - Highest signal weekly AI news show is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.OpenAI drops GPT 5.4 Thinking and 5.4 Pro - heavy weight frontier models with 1M context, /fast mode, SOTA on many evalsOpenAI actually opened this week with another model drop, GPT 5.3-instant, which... we can honestly skip, it was fairly insignificant besides noting that this is the model that most free users use. It is supposedly “less cringe” (actual words OpenAI used). We all wondered when 5.4 will, and OpenAI once again proved that we named the show after the right day. Of course it drops on a ThursdAI. GPT 5.4 Thinking is OpenAI latest “General” model, which can still code, yes (they folded most of the Codex 5.3 coding breakthroughs in here) but it also shows an incredible 83% on GDPVal (12% over Codex), 47% on Frontier Math and an incredible ability to use computers and browsers with 82% on BrowseComp beating Claude 4.6 at lower prices than Sonnet! GPT 5.4 is also ... quite significantly improved at Frontend design? This landing page was created by GPT 5.4 (inside the Codex app, newly available on Windows) in a few minutes, clearly showing significant improvements in style. I built it also to compare prices, all the 3 flagship models are trying to catch up to Gemini in 1M context window, and it's important to note, that GPT 5.4 even at double the price after the 272K tokens cutoff is still.... cheaper than Opus 4.6. OpenAI is really going for broke here, specifically as many enterprises are adopting Anthropic at a faster and faster pace (it was reported that Anthropic is approaching 19B ARR this month, doubling from 8B just a few months ago!) Frontier math wizThe highlight from the 5.4 feedback came from a Polish mathematician Bartosz Naskręcki (@nasqret on X), who said GPT-5.4 solved a research-level FrontierMath problem he had been working on for roughly 20 years. He called it his “personal singularity,” and as overused as that word has become, I get why he said it. I've told you about this last week, we're on the cusp. Coding efficiencyThere's tons of metrics in this release, but I wanted to highlight this one, where it may seem on first glance that on SWE-bench Pro, this model is on par with the previous SOTA GPT 5.3 codex, but these dots here are thinking efforts. And a medium thinking effort, GPT 5.4 matches 5.3 on hard thinking! This is quite remarkable, as lower thinking efforts have less tokens, which means they are cheaper and faster ultimately! Fast mode arrives at OpenAI as wellI think this one is a direct “this worked for Anthropic, lets steal this”, OpenAI enabled /fast mode that.. burns the tokens at 2x the rate, and prioritizes your tokens at 1.5x the speed. So, essentially getting you responses faster (which was one of the main complains about GPT 5.3 Codex). I can't wait to bring the fast mode to OpenClaw with 5.4, which will absolutely come as OpenClaw is part of OpenAI now. There's also a really under-appreciated feature here that I think other labs are going to copy quickly: mid-thought steering. OpenAI now lets you interrupt the model while it's thinking and redirect it in real time in ChatGPT and iOS. This is a godsend if you're like me, sent a prompt, seeing the model go down the wrong path in thinking... and want to just.. steer it without stopping! Anthropic is now designated as supply-chain risk by DoWLast week I left you with a cliffhanger: Anthropic had received an ultimatum from the Department of War (previously the Department of Defense) to remove their two remaining restrictions on Claude — no autonomous kill chain without human intervention, and no surveillance of US citizens. Anthropic's response? “we cannot in good conscience acceede to their request” So much has happened since then; US President Trump said “I fired Anthropic” referring to his Truth Social post demanding intelligence agencies drop the use of Claude (which apparently was used in the war with Iran regardless); Sam Altman announced that OpenAI has agreed to DoW and will provide OpenAI models, causing a lot of people to cancel their OpenAI subscriptions, and later apologizing for the “rushed rollout”; Dario Amodei posted a very contentious internal memo that leaked, in which he name-called the presidency, Sam Altman and his motives, Palantir and their “safety theater”, for which he later apologizedHonestly this whole thing is giving me whiplash trying to follow, but here's the facts. Anthropic is now the first US company in history, being designated “supply chain risk” which means no government agency can use Claude, and neither can any company that does contracts with DoW. Anthropic says it's illegal and will challenge this in court , while reporting $19B in annual recurring revenue, nearly doubling since last 3 months, and very closely approaching OpenAI at $25B. Look, did I want to report on this stuff when I decided to cover AI? no... I wanted to tell you about cool models and capabilities, but the world is changing, and it's important to know that the US Government understands now that AI is inevitable, and I think this is just the first of many clashes between tech and government we'll see. We'll keep reporting on both. (but let me know in the comments if you'd prefer just model releases) OpenAI's GPT-5.3 Instant Gets Less Cringe, Google's Flash-Lite Gets Faster (X, Announcement)We also got two fast-model updates this week that are worth calling out because these are the models that often end up powering real product flows behind the scenes. As I wrote before, OpenAI's instant model is nothing to really mention, but it's worth mentioning that OpenAI seems to have an answer for every Gemini release. Gemini released Gemini Flash-lite this week, which boasts an incredible 363 tokens/s speed, which doing math at a very good level, 1M context and great scores compared to the instant/fast models like Haiku from Anthropic. Folks called out that this model is more expensive than the previous 2.5 Flash-lite. But with 86.9% on GPQA Diamond beating GPT-5 mini, and 76.8% MMMU-pro multimodal reasoning, this is definitely worth taking a look at for many agentic, super fast responses! For example, the heartbeat response in OpenClaw. Qwen 3.5 Small Models & The Departure of Junyang Lin (X, HF, HF, HF)Alibaba's Qwen team continued releasing their Qwen 3.5 family, this time with Qwen 3.5 Small, a series of models at 0.8B, 2B, 4B, and 9B parameters with native multimodal capabilities. The flagship 9B model is beating GPT-OSS-120B on multiple benchmarks, scoring 82.5 on MMLU-Pro and 81.7 on GPQA Diamond. These models can handle video, documents, and images natively, support up to 201 languages, and can process up to 262K tokens of context. And.. they are great! They are trending on HF right now. What's also trending is, tech lead for Qwen, a friend of the pod Junyang Lin, has posted a cryptic tweet that went viral with over 6M views. There was a lot of discussions on why he and other Qwen leads are stepping away, what's goig to happen with the future of OpenSource. The full picture seems to be, there are a lot of internal tensions and politics, with Junyang being one of the youngest P10 leaders in the Alibaba org.A Chinese website 36KR ( Kind of like a chinese techcrunch) reported that this matter went all the way up to Alibaba CEO, who is no co-leading the qwen team, and that this resignation was related to an internal dispute over resource allocation and team consolidation, not a firing. I'm sure Junyang is going to land somewhere incredible and just wanted to highlight just how much he did for the open source community, pushing Qwen relentlessly, supporting and working with a lot of inference providers (and almost becoming a co-host for ThursdAI with 9! appearances!) StepFun releases Step 3.5 Flash Base (X, HF, HF, Announcement, Arxiv)Speaking of Open Source, StepFun just broke through the noise with a new model, a 196B parameter sparse Mixture of Experts model activating just 11B parameters when ran. It has some great benchmarks, but the main thing is this: they are releasing the pretrained base weights, a midtrain checkpoint optimized for code and agents, the complete SteptronOSS training framework, AND promising to release their SFT data soon - all under Apache 2.0! Technically the model looks strong too, with multi-token prediction, 74.4% on SWE-bench verified bench (though, as we told you last week, it's.. no longer trusted) and full apache 2! This Week's Buzz: presenting Wolfbench.ai I'm so excited about this weeks “this weeks buzz”, Wolfram has been hard at work preparing and presenting a new framework to test out these models, and named it wolfbench.ai Wolfbench is our attempt to compare how the same model performs via different agentic harnesses like ClaudeCode, OpenClaw and Terminalbench's own Terminus. You can check out the website on wolfbench.com but the short of it is, a single number is not telling the full story. Wolf Bench breaks it into a four-metric framework: the average score across runs, the best single run, the ceiling (how many tasks can the model solve at least once across all runs), and the floor (how many tasks does it solve consistently across every single run). That last one is what I find most illuminating. Opus 4.6 might be able to solve 88% of Terminal Bench tasks on average, but only about 55% of tasks it solves every single time. Reliability matters enormously for agents, and benchmarks almost never surface this. If you want to run your own evals with the same config, reach out to Wolfram—he's open to community contributions. Wolfram has also already kicked off a Wolf Bench run on GPT-5.4 since we tested it live today, so stay tuned for those results.There's quite a few more releases we didn't have time to get into on the show given the GPT 5.4 drop, you'll find all those links in the show notes! Next week will mark 3 years since I've started talking about AI on the internet and created ThursdAI (It was March 14th, 2023, same day as GPT4 launched) and we'll have a little celebration, I do hope you join us live
On this episode, we catch up with Julia Castro (@fuertejulia), the versatile Spanish waterwoman from Fuerteventura, during her time repping Foil Drive at BOOT Düsseldorf 2026. Fresh from the massive indoor pool demos in "the surf hole" at the world's leading water sports trade show, Julia shares the electric vibe of the event, her journey into foil assist tech, and candid thoughts on the evolving state of professional watersports.Episode Highlights:Life inside BOOT Düsseldorf: 5 days deep in the massive yachting halls, repping Foil Drive in the dedicated surf/foil area with a huge indoor pool setup—constant crowds, endless questions, and non-stop demos showing why foil assist is exploding in popularityHow Julia discovered Foil Drive by chance at her local spot, went from skeptic ("this cannot be true, this is sorcery") to instant addict after binge-watching videos, and now handles marketing, content, and pro team duties for Foil Drive Europe after nearly two yearsThe groundbreaking side of Foil Drive: first brand to make a truly universal mast-mount motor, plus their pioneering collaborative spirit—openly partnering with Slingshot, Armstrong, F-One, Axis, and more to grow the whole foil ecosystem instead of gatekeepingOvercoming major personal hurdles: shoulder destruction, a cancer scare right after recovery, COVID lockdowns in strict Spain, sponsor cuts, and shifting from full-time competition to event filming/social media work (now 60-70% of her income) while staying positive and healthyThe tough realities of modern watersports sponsorship: the shift from skill-based deals to needing massive social media fame, pressure on guys to go ultra-extreme (with scary injuries), and the unfortunate bikini-heavy expectations for women—Julia fights for real choice, authenticity, and legacy beyond "only fans" style contentWhy brands hold huge responsibility: pushing real people as role models instead of curated illusions that harm mental health (especially kids idolizing non-real personas), prioritizing quality/value over raw views, and supporting diverse paths for women in the sportFoil Drive fun and future: addictive for water lovers, game-changing for wave sessions (catching 30-50 waves vs. friends' 5), altered dynamics (more rear-foot weight), potential SFT foil-assist/wave divisions, and her dream to compete again—plus a cheeky plan to "borrow" a battery for a van road trip back to Fuerteventura with snowboarding and foil stopsGrowing up in paradise: Fuerteventura childhood on the beach, late start to kiting/watersports (tourist-priced), realizing after world travel that home really is one of the best spots, now packed with visitorsIf you're into foil assist tech like Foil Drive, the intersection of e-foiling/wave foiling/pump foiling, honest talks on gender dynamics and mental health in action sports, athlete resilience stories, or just pure stoke from someone living the waterwoman life—this episode delivers real talk, inspiration, and plenty of foil obsession!Catch the full conversation and follow Julia Castro on Instagram @fuertejulia for her road trip adventures, Foil Drive sessions, event coverage, and more. Big thanks to Foil Drive for the ongoing innovation, and stay tuned for more from the growing foil world in 2026. Listen now!
What's really going on in food and farming? Two farmers – Patrick Holden, founder of the Sustainable Food Trust, and Stuart Oates, founder of the Fossil Free Farm project – get behind the headlines to unpack the biggest stories shaping what we eat, how we farm and the future of the planet. Expect lively debate, real-world experience, and unique insights from some of food and farming's top voices. To mark the release of the recent documentary Finding Harmony: A King's Vision – exploring his Majesty King Charles III's lifelong commitment to the environment and his philosophy of Harmony – in this episode of the SFT Podcast, Patrick talks to Stuart about his involvement in the documentary and recalls first meeting the King in 1982. They also explore the philosophy of Harmony in more depth, including how we can bring harmonious practices into our everyday lives and why achieving this is now more important than ever. Elsewhere in the episode, Patrick and Stuart talk about the evolution of agricultural education and the importance of influencing the influencers to enable a transition to more sustainable food and farming practices. To join in the conversation, get in touch with us at info@sustainablefoodtrust.org or send us a message via any of our social media channels – let us know what you'd like to hear Patrick and Stuart talk about next time! Timestamps: 0:00: Welcome back! 0:33: Patrick talks about his relationship with King Charles, first meeting him in 1982, the development of Highgrove Farm and their shared passion for organic farming 4:01: The King's book: Harmony: A New Way of Looking at Our World 5:55: We are not separate from nature, we are part of nature 10:45: Striving for Harmony in a world of conflict 13:36: To achieve Harmony, where do we start? 18:02: Harmonious practices in our everyday lives 25:48: Patrick's involvement in Finding Harmony 28:36: 'Seeing is believing' experience at Highgrove Farm 30:17: The evolution of agricultural education and need for reform 37:53: The history of Dumfries House 43:14: The interconnectedness of faiths/The King's influence on connecting different faiths 44:52: How do we transform farming to become part of the solution, rather than the problem? 49:35: Influencing the influencers 53:31: What's coming in future episodes Follow Patrick and the Sustainable Food Trust: https://www.instagram.com/susfoodtrust/ | https://www.instagram.com/hafodcheese/ https://bsky.app/profile/susfoodtrust.bsky.social https://www.facebook.com/SusFoodTrust https://www.linkedin.com/company/sustainable-food-trust/ Find out more about Stuart and the Fossil Free Farm project: https://fossilfreefarm.com/ https://www.instagram.com/camelstu/ https://www.youtube.com/@farmerstu To listen to more SFT podcasts, featuring some of the biggest names in regenerative food and farming, head to our main podcast page.
If you listened to my last episode about creating income by teaching without becoming an influencer and thought, "That sounds great… but when would I actually find the time?",this episode is for you. In this conversation, I talk honestly about time management for therapists who want to create continuing education, without hustle culture, productivity pressure, or burnout. I break down why "I don't have time" makes complete sense for clinicians, and why the real issue usually isn't time at all, it's structure, energy, and permission. I explore the difference between clinical energy and creative energy, why waiting for long blocks of free time keeps therapists stuck, and how continuing education can be built in small, realistic containers that actually fit therapist life. I also share why most therapists already have far more teachable material than they realize and why starting smaller than you think is often the key to getting unstuck. This episode is especially for therapists who feel curious about teaching or creating CE, but overwhelmed by where it would fit, or whether they're "ready" to begin. Special Promotion: Berries + Free CE Podcourse Bundle (New Annual Customers) At the beginning of this episode, I shared a limited-time promotion in partnership with Berries. New customers who purchase the Berries annual subscription plan using my referral link will receive free access to my CE Podcourse Bundle, which includes over 30 hours of NBCC-approved continuing education contact hours, with new podcourses added throughout the year. A Podcourse is a podcast and an audio course in one - designed for busy clinicians. You can listen on a walk, between sessions, or whenever it fits into your schedule. When you're ready, you log in, complete a short self-study quiz, and download your certificate of completion.
Many therapists feel a pull to create income outside of the one-to-one therapy model, but feel conflicted about influencer culture, constant posting, or anything that feels misaligned with their professional values. In this episode, I explore how therapists can diversify their income by teaching, without becoming influencers or leaving the field. I break down why teaching is a natural extension of clinical work, how continuing education fits ethically within scope, and why so many therapists already have teachable expertise without realizing it. I discuss practical, realistic options for therapists, including continuing education, consultation, digital educational products, and podcast-based learning, and reframe income diversification as an issue of sustainability, not commitment. This episode is especially relevant for seasoned clinicians who want longevity in the profession without hype or hustle culture. Special Promotion: Berries + Free CE Podcourse Bundle (New Annual Customers) At the beginning of this episode, I shared a limited-time promotion in partnership with Berries. New customers who purchase the Berries annual subscription plan using my referral link will receive free access to my CE Podcourse Bundle, which includes over 30 hours of NBCC-approved continuing education contact hours, with new podcourses added throughout the year. A Podcourse is a podcast and an audio course in one - designed for busy clinicians. You can listen on a walk, between sessions, or whenever it fits into your schedule. When you're ready, you log in, complete a short self-study quiz, and download your certificate of completion.
Improve your foiling skills in paradise! Join us in Montanita Ecuador May 23-30, 2026 for a foil drive / tow / prone foil camp with Ecuador Foil, KT Foiling & Julia Castro. Learn MoreOn this episode we sit down with Edan Fiander, the young Swiss powerhouse and reigning SFT Pump Foil World Champion. Fresh off his dominant win at the BOOT Düsseldorf event to kick off the 2026 season, Edan opens up about his rapid rise in the emerging world of competitive pump foiling—from his roots on Lake Geneva to claiming multiple Swiss titles and the global crown.Episode Highlights:- Edan's journey from a skateboarding background (with plenty of injuries) to discovering pump foiling as the perfect no-wind, no-wave solution for Switzerland's Lake Geneva—turning a casual try into daily obsession and a full community at Tropical Corner with around 250 yearly pass holders- Why pump foiling stands out as one of the toughest foiling disciplines to start (needing pure self-generated power and balance) yet offers endless accessibility—no wind or waves required—and cross-training benefits for wing foiling, surf foiling, and beyond- Training secrets behind his success: consistent dock starts, explosive full-range strength work, strapless sessions for feel, high-level coaching (physio, massages, and prep), altitude mask breathing drills for respiratory power, and handling turbulence from other riders in tight races- Competing in the SFT: from a rushed debut in Düsseldorf (finishing 5th after a foil issue) to winning the event this year; key factors like lightning-fast starts, energy management, risk calculation in heats, mindset, and avoiding common falls (toe slips in straps, speed wobbles, buoy touches)- Gear insights: riding the Lift HA 120 (775 cm², ~10.2 aspect ratio) for its perfect balance of glide, speed, playfulness, and tight turns—ideal for technical SFT courses—plus how modern high-aspect foils have evolved for racing vs. all-around use- Life as a 19-year-old multi-time champion: balancing a gap year of intense training with upcoming university studies in business management, staying motivated to defend his title, and the joy of growing a small but passionate sport- Bonus vibes: Swiss community spirit, why pump foiling opens doors for flat-water riders everywhere, quick tips for aspiring competitors, and even a taste of Swiss desserts (meringue with double cream from Gruyère—rich, not healthy!)If you're into pump foiling, competitive foiling scenes, training deep dives, gear geekery, or stories of young athletes pushing limits in non-wind-powered sports—this episode is packed with motivation, technique breakdowns, and pure stoke!Catch the full conversation and stay tuned to the Surf Foil World Tour (SFT) for more pump foil action in 2026.Follow Edan Fiander on Instagram @edan.fndr for clips, training, and updates, and check out the Foil Life Podcast channels for the episode drop. Listen now!
On this episode we sit down with Edan Fiander, the young Swiss powerhouse and reigning SFT Pump Foil World Champion. Fresh off his dominant win at the BOOT Düsseldorf event to kick off the 2026 season, Edan opens up about his rapid rise in the emerging world of competitive pump foiling—from his roots on Lake Geneva to claiming multiple Swiss titles and the global crown.Episode Highlights:- Edan's journey from a skateboarding background (with plenty of injuries) to discovering pump foiling as the perfect no-wind, no-wave solution for Switzerland's Lake Geneva—turning a casual try into daily obsession and a full community at Tropical Corner with around 250 yearly pass holders- Why pump foiling stands out as one of the toughest foiling disciplines to start (needing pure self-generated power and balance) yet offers endless accessibility—no wind or waves required—and cross-training benefits for wing foiling, surf foiling, and beyond- Training secrets behind his success: consistent dock starts, explosive full-range strength work, strapless sessions for feel, high-level coaching (physio, massages, and prep), altitude mask breathing drills for respiratory power, and handling turbulence from other riders in tight races- Competing in the SFT: from a rushed debut in Düsseldorf (finishing 5th after a foil issue) to winning the event this year; key factors like lightning-fast starts, energy management, risk calculation in heats, mindset, and avoiding common falls (toe slips in straps, speed wobbles, buoy touches)- Gear insights: riding the Lift HA 120 (775 cm², ~10.2 aspect ratio) for its perfect balance of glide, speed, playfulness, and tight turns—ideal for technical SFT courses—plus how modern high-aspect foils have evolved for racing vs. all-around use- Life as a 19-year-old multi-time champion: balancing a gap year of intense training with upcoming university studies in business management, staying motivated to defend his title, and the joy of growing a small but passionate sport- Bonus vibes: Swiss community spirit, why pump foiling opens doors for flat-water riders everywhere, quick tips for aspiring competitors, and even a taste of Swiss desserts (meringue with double cream from Gruyère—rich, not healthy!)If you're into pump foiling, competitive foiling scenes, training deep dives, gear geekery, or stories of young athletes pushing limits in non-wind-powered sports—this episode is packed with motivation, technique breakdowns, and pure stoke!Catch the full conversation and stay tuned to the Surf Foil World Tour (SFT) for more pump foil action in 2026.Follow Edan Fiander on Instagram @edan.fndr for clips, training, and updates, and check out the Foil Life Podcast channels for the episode drop. Listen now!
Sir Julian Rose and his daughter, Miriam Rose, of the Hardwick Estate – England's first community-owned estate – joined our CEO, Patrick Holden, on the SFT Podcast. Julian Rose was one of the pioneers of organic farming in the UK. In the late 1960s, Julian inherited ownership of the Hardwick Estate and by 1975 the 375-acre farmland had been converted to organic status. 40 years later, the Rose family have made the decision to transfer ownership of the Estate to the local community, seeking to move away from a history of private aristocratic ownership. Following closely in her father's footsteps, Miriam says she identified as a 'natural environmentalist' from a very young age having grown up on the farm, and has been a fierce advocate ever since – including spending eight days incarcerated in Iceland after protesting the country's aluminium industry's impact on the rivers. Nowadays, Miriam sits on the board of Trustees of the Hardwick Estate, overseeing its direction, which is otherwise managed by the Hardwick community. In this episode, Julian and Miriam go into more detail about how this community-owned model works, as well as the risks involved with this approach, and they discuss whether or not this model could be scaled. They also talk about the complex feelings they have towards land-ownership and the inheritance of wealth. Find out more about the Hardwick Estate here. This conversation was recorded in October 2025. Featured image courtesy of www.julianrose.info/. Timestamps 0:00: Welcome Julian and Miriam! 01:04: The history of the Hardwick Estate and how the Julian came to inherit the Estate 9:05: Patrick and Julian's involvement in the early organic movement – including the beginnings of the UK's organic food standards 12:15: Miriam's early beginnings as an environmentalist and her role at Hardwick 15:12: Becoming England's first community-owned estate 22:35: How does the community ownership model work? 24:46: What are the risks involved with this approach? 30:02: 50 years of organic at the Hardwick Estate 32:58: Can this community-owned model be scaled? 33:59: The complexities of inheriting land and wealth To listen to more SFT podcasts, featuring some of the biggest names in regenerative food and farming, head to our main podcast page. And to keep up to date with our news, you can subscribe to our monthly newsletter or follow us on Instagram, Facebook, LinkedIn and Bluesky.
In this episode of The Therapy Show, I chat with Ann Mac Prevost, a licensed professional counselor who specializes in teen anxiety. Ann Mac dives into how anxiety shows up in today's teens, why it's more prevalent than ever, and how parents and therapists can help teens manage it effectively. We also explore the role of family systems, the impact of social media, and practical tools that support emotional growth in adolescents. In this episode, we cover: The most common anxiety presentations in teens today Why COVID and smartphones have intensified anxiety in adolescents How to know when anxiety crosses from "normal" to "problematic" Tips for therapists working with teens and their families The importance of involving parents in the therapeutic process Helpful frameworks for validating emotions while promoting behavior change Practical CBT and exposure strategies that teens can actually use Whether you're a therapist, parent, or just curious about teen mental health, this conversation is packed with relatable insights and tools you can use right away. Connect with Ann Mac. Links mentioned:
At some point in your career as a therapist, the work begins to shift. You may feel a quiet pull to share what you've learned beyond the therapy room but feel unsure what that means or whether you're "qualified" to teach. In this episode, I explore the transition from clinician to educator and why this identity shift can feel both exciting and uncomfortable. I talk about common myths therapists hold about teaching, ethical considerations, visibility, and how stepping into an educator role doesn't require a new certification or a major career change. If you're a mid-career or seasoned clinician curious about teaching, mentoring, or continuing education, this episode is an invitation to notice the pull and explore it with clarity and confidence. Links mentioned:
Improve your foiling skills in paradise! Join us in Montanita Ecuador May 23-30, 2026 for a foil drive / tow / prone foil camp with Ecuador Foil, KT Foiling & Julia Castro. Learn MoreOn this episode, we're diving into the thrilling world of SFT e-foiling with Agnes Wicander — the Swedish powerhouse, Waydoo ambassador, sales & marketing rep, and standout women's SFT e-foil champion who dominated the scene in her debut racing season!Episode Highlights:- Agnes's journey from a sporty Swedish family (kite foiling since 2015, yachting background) to becoming a top e-foil racer — spotting the Waydoo Kickstarter, starting with her dad in Sweden, touring Europe in a 6-meter camper van to build the brand, and shifting from chill cruising to high-speed, high-adrenaline racing- Why e-foiling is the perfect complement to wind sports (wing foiling, kiting, paragliding) — ultimate accessibility in low-wind Sweden, tech appeal, silent exploration, and that addictive “foil throttle” rush that feels more like driving a fast car than anything else on the water- Behind-the-scenes of competitive e-foiling: inaugural SFT & E-Foil Racing League seasons, technical courses rewarding skill over pure speed, prop wash battles, nerve management in sensitive-trigger racing, common falls (sharp turns & wing breaches), evolving rules, and how racing pushes brands like Waydoo to innovate (lighter batteries, better wings, upcoming Foil Boost for swell & wave riding)- Gear talk & setup tweaks: Shimming madness (2 degrees too much = no lift!), custom wing choices for different riding styles, reverse mode for board recovery, GPS tracking, challenges in chop/swell without motor glide, and the dream of wing-breach capable foils- Epic e-foiling destinations unlocked by wind independence: ultra-remote Papua New Guinea sugar-top islands, serene Norwegian fjords, canal cruising in Australia (coffee run on the board!), backcountry lakes, camping adventures, and exploring places a boat or jet ski could never reach- The supportive women's division vibe, growing female participation, ideas for making racing more accessible (battery rentals, event support), and why e-foiling delivers a unique adrenaline hit that's hard, technical, and seriously addictiveIf you're curious about motorized foiling, the future of e-foil racing, silent exploration in any conditions, the massive progression from cruiser to competitor, or just want to feel that pure throttle rush — this episode is packed with inspiration and stoke!Listen to the full conversation with Agnes on the Foil Life Podcast channels. Follow Agnes on instagram @agneswicander
This episode is brought to you by Armstrong Foils. Armstrong Foils are a founding member and proud supporter of the all new Global Foil Board Sports Association (GFA) To learn more visit: www.gfafoilworld.com / www.armstrongfoils.comOn this episode, we're diving into the thrilling world of SFT e-foiling with Agnes Wicander — the Swedish powerhouse, Waydoo ambassador, sales & marketing rep, and standout women's SFT e-foil champion who dominated the scene in her debut racing season!Episode Highlights:- Agnes's journey from a sporty Swedish family (kite foiling since 2015, yachting background) to becoming a top e-foil racer — spotting the Waydoo Kickstarter, starting with her dad in Sweden, touring Europe in a 6-meter camper van to build the brand, and shifting from chill cruising to high-speed, high-adrenaline racing- Why e-foiling is the perfect complement to wind sports (wing foiling, kiting, paragliding) — ultimate accessibility in low-wind Sweden, tech appeal, silent exploration, and that addictive “foil throttle” rush that feels more like driving a fast car than anything else on the water- Behind-the-scenes of competitive e-foiling: inaugural SFT & E-Foil Racing League seasons, technical courses rewarding skill over pure speed, prop wash battles, nerve management in sensitive-trigger racing, common falls (sharp turns & wing breaches), evolving rules, and how racing pushes brands like Waydoo to innovate (lighter batteries, better wings, upcoming Foil Boost for swell & wave riding)- Gear talk & setup tweaks: Shimming madness (2 degrees too much = no lift!), custom wing choices for different riding styles, reverse mode for board recovery, GPS tracking, challenges in chop/swell without motor glide, and the dream of wing-breach capable foils- Epic e-foiling destinations unlocked by wind independence: ultra-remote Papua New Guinea sugar-top islands, serene Norwegian fjords, canal cruising in Australia (coffee run on the board!), backcountry lakes, camping adventures, and exploring places a boat or jet ski could never reach- The supportive women's division vibe, growing female participation, ideas for making racing more accessible (battery rentals, event support), and why e-foiling delivers a unique adrenaline hit that's hard, technical, and seriously addictiveIf you're curious about motorized foiling, the future of e-foil racing, silent exploration in any conditions, the massive progression from cruiser to competitor, or just want to feel that pure throttle rush — this episode is packed with inspiration and stoke!Listen to the full conversation with Agnes on the Foil Life Podcast channels. Follow Agnes on instagram @agneswicander
In this episode I had a powerful conversation with Dr. Andrew Hartz, a clinical psychologist and founder of the Open Therapy Institute. We explored how politics, values, and therapy are increasingly showing up together in the therapy room and why that matters for both clients and clinicians. What's New with Berries: Berries now lets you generate a complete, personalized treatment plan in seconds - built from your diagnoses, session notes, and clinical preferences, using customizable templates that match your voice and style. Its powerful "golden thread" ensures your treatment plan and notes stay clinically aligned, continuously informing each other as care evolves. With the new Magic Update feature, your plan updates effortlessly without the need to rewrite from scratch. The result? A streamlined workflow where every session builds on the last, and documentation becomes part of your clinical process - not just another admin task. Use code TherapyShow50 for $50 off your first month - CLICK HERE. Key takeaways: Therapist bias is a real and growing concern. Many clients feel alienated due to perceived political or ideological leanings of their therapists, often unintentionally communicated through things like pronoun usage or assumptions about worldview. Most therapists lean left politically, which can lead to clients self-censoring, feeling misunderstood, or avoiding therapy altogether. The Open Therapy Institute (OTI) was created to support therapists who want to offer politically neutral, values-attuned therapy and serve populations that feel underserved, especially those with conservative or centrist views. Therapists can grow their practice by learning to work effectively with clients from across the political spectrum. There is high demand and low supply of therapists trained to do this well. We discussed the importance of dialectical thinking. This means helping clients (and ourselves) hold multiple perspectives and tolerate ambiguity, especially around politics, religion, and identity. If you're a therapist who wants to grow in this area or reach more clients who feel left out by traditional therapy, check out Open Therapy Institute, https://opentherapyinstitute.org. Browse all my NBCC approved Podcourses - just $5 each. Get one CE contact hour. Build your first CE course (free) Get my Coping with Political Stress Ebook and Peaceful Politics AI Guide Therapist Conversation Framework: Politics in Session A printable PDF with 97 questions to navigate political talk in therapy - without taking sides. Solution-Focused Therapy Guide72 questions + prompts to help adult clients clarify goals and move forward using SFT. Check out all my Counselor Resources.
In this episode of The Effortless Podcast, Amit Prakash and Dheeraj Pandey are joined by Alex Dimakis for a wide-ranging, systems-first discussion on the future of long-horizon AI agents that can operate over time, learn from feedback, adapt to users, and function reliably inside real-world environments.The conversation spans research and industry, unpacking why prompt engineering alone collapses at scale; how advisor models, reward-driven learning, and environment-based evaluation enable continual improvement without retraining frontier models; and why memory in AI systems is as much about forgetting as it is about recall. Drawing from distributed systems, reinforcement learning, and cognitive science, the trio explores how personalization, benchmarks, and context engineering are becoming the foundation of AI-native software.Alex, Dheeraj, and Amit also examine the evolution from SFT to RL to JEPA-style world models, the role of harnesses and benchmarks in measuring real progress, and why enterprise AI has moved decisively from research into engineering. The result is a candid, deeply technical conversation about what it will actually take to move beyond demos and build agents that work over long horizons.Key Topics & Timestamps 00:00 – Introduction, context, and holiday catch-up04:00 – Teaching in the age of AI and why cognitive “exercise” still matters08:00 – Industry sentiment: fear, trust, and skepticism around LLMs12:00 – Memory in AI systems: documents, transcripts, and limits of recall17:00 – Why forgetting is a feature, not a bug22:00 – Advisor models and dynamic prompt augmentation27:00 – Data vs metadata: control planes vs data planes in AI systems32:00 – Personalization, rewards, and learning user preferences implicitly37:00 – Why prompt-only workflows break down at scale41:00 – RAG, advice, and moving beyond retrieval-centric systems46:00 – Long-horizon agents and the limits of reflection-based prompting51:00 – Environments, rewards, and agent-centric evaluation56:00 – From Q&A benchmarks to agents that act in the world1:01:00 – Terminal Bench, harnesses, and measuring real agent progress1:06:00 – Frontier labs, open source, and the pace of change1:11:00 – Context engineering as infrastructure (“the train tracks” analogy)1:16:00 – Organizing agents: permissions, visibility, and enterprise structure1:20:00 – SFT vs RL: imitation first, reinforcement last1:25:00 – Anti-fragility, trial-and-error, and unsolved problems in continual learning1:28:00 – Closing reflections on the future of long-horizon AI agentsHosts:Amit PrakashCEO & Founder at AmpUp, Former engineer at Google AdSense and Microsoft Bing, with deep expertise in distributed systems, data platforms, and machine learning.Dheeraj PandeyCo-founder & CEO at DevRev, Former Co-founder & CEO of Nutanix. A systems thinker and product visionary focused on AI, software architecture, and the future of work.Guest:Alex DimakisAlex Dimakis is a Professor in UC Berkeley in the EECS department. He received his Ph.D. from UC Berkeley and the Diploma degree from NTU in Athens, Greece. He has published more than 150 papers and received several awards including the James Massey Award, NSF Career, a Google research award, the UC Berkeley Eli Jury dissertation award, and several best paper awards. He is an IEEE Fellow for contributions to distributed coding and learning. His research interests include Generative AI, Information Theory and Machine Learning. He co-founded Bespoke Labs, a startup focusing on data curation for specialized agents.Follow the Hosts and the Guest: Dheeraj Pandey:LinkedIn - https://www.linkedin.com/in/dpandeyTwitter - https://x.com/dheerajAmit Prakash:LinkedIn - https://www.linkedin.com/in/amit-prak...Twitter - https://x.com/amitp42Alex Dimakis:LinkedIn - https://www.linkedin.com/in/alex-dima...Twitter - https://x.com/AlexGDimakis Share Your Thoughts Have questions, comments, or ideas for future episodes?
Our CEO, Patrick Holden, recently sat down for a conversation with Sir Tim Smit, former archaeologist, music producer and founder of the Eden Project and Lost Gardens of Heligan in Cornwall. In this episode, Patrick and Tim talk about the emerging dangers of forever chemicals for the environment and human health, reconceptualising what we mean when we talk about farms and their importance for mental health and wellbeing, and why the integration of food and farming into the national curriculum "shouldn't be fringe, it should be right at the core". To listen to more SFT podcasts, featuring some of the biggest names in regenerative food and farming, head to our main podcast page. And to keep up to date with our news, you can subscribe to our monthly newsletter or follow us on Instagram, Facebook, LinkedIn and Bluesky. This conversation was recorded in December 2025.
I'm breaking down what to do when the Facebook group grind stops working. I share real, practical strategies that I've used myself to market my continuing education Podcourses, without burning out or relying solely on social media. I talk about: Why Facebook posts lose momentum How to build a simple email marketing funnel that actually works Why partnerships and podcast guesting can open new doors The power of searchable, evergreen content And how to make it ridiculously easy for someone to buy your course If you're tired of spinning your wheels, this episode will help you pivot, not panic. Plus, I give you a behind-the-scenes look at how I market my Podcourse bundle, including what didn't work and what finally clicked. Read the blog here. Links mentioned in this episode: Browse all the Podcourses Build your first CE course (free) Save time with Berries AI: get $50 off your first month with code THERAPYSHOW50 Get my Coping with Political Stress Ebook and Peaceful Politics AI Guide Therapist Conversation Framework: Politics in Session A printable PDF with 97 questions to navigate political talk in therapy - without taking sides. Solution-Focused Therapy Guide72 questions + prompts to help adult clients clarify goals and move forward using SFT. Check out all my Counselor Resources.
I'm helping you plan your 2026 growth goals in this episode, whether that's continuing education courses, certifications, clinical skills, or supervision. If you're feeling a little unsure about where to focus next year, this is your chance to reflect, regroup, and set a clear, simple direction for your professional development. I'll walk you through how to identify what actually worked in 2025, choose what's worth your time in 2026, and share a few of my favorite tools that save time and reduce overwhelm - including a note-writing AI I'm loving, and a free CE Course Builder I created just for therapists like you. In this episode, I cover: How to plan CEUs you'll actually enjoy Choosing certifications that align with your goals Building your skills and setting one focus per quarter Tools to streamline your work and free up space to grow Links mentioned in this episode: Browse all the Podcourses Build your first CE course (free)Save time with Berries AI: get $50 off your first month with code THERAPYSHOW50 Get my Coping with Political Stress Ebook and Peaceful Politics AI Guide Therapist Conversation Framework: Politics in Session A printable PDF with 97 questions to navigate political talk in therapy - without taking sides. Solution-Focused Therapy Guide72 questions + prompts to help adult clients clarify goals and move forward using SFT. Check out all my Counselor Resources.
Sponsored by Berries AI: Use code TherapyShow50 for $50 off your first month - CLICK HERE. If you are a therapist or counselor looking for continuing education, check out my NBCC Approved $5 Podcourses and other continuing education offerings. Plus, get your first Podcourse half off. In this episode of The Therapy Show, I share something I've been working on behind the scenes - a free tool I created just for mental health clinicians: the CE Course Builder, a custom GPT designed to help you create and launch your own continuing education courses. If you've ever thought about teaching but felt overwhelmed by the tech, compliance, or where to even start, this tool walks you through it all - step-by-step. I also talk about group discounts available for practice owners (email me to discuss offering my CE Podcourses to your clinicians) and invite you to fill out a short survey to help shape future CE content. If you're ready to move beyond the therapy room and share your expertise, this episode is for you. Get my Coping with Political Stress Ebook and Peaceful Politics AI Guide Therapist Conversation Framework: Politics in Session A printable PDF with 97 questions to navigate political talk in therapy - without taking sides. Solution-Focused Therapy Guide72 questions + prompts to help adult clients clarify goals and move forward using SFT. Check out all my Counselor Resources.
This episode is brought to you by Villa Carina Apartments in beautiful Bonaire. In this episode, we sit down with the newly crowned 2025 E-Foil Surf Foil World Tour (SFT) World Champion – the undefeated e-foil racer who took the title in the season finale in Abu Dhabi.Fresh off dominating the inaugural SFT season, the Florida-based ripper (and Flightboard early adopter) joins us to break down what it actually feels like to turn a five-year hunch into a world championship, how e-foil racing went from “nice idea” to a full-blown global tour in record time, and why this sport is exploding faster than anyone predicted.We go deep on:- From kite-smash accidents to building one of the first e-foil schools in South Florida - The wild Atlanta Foil Fest Enduro with Brian Grubb, Nick Leeson, and 20 riders dodging submerged trees at full throttle - Unsanctioned full-send dawn patrols through Amsterdam's canals (don't try this at home) - Gear geek-out: custom shims, chopped tails, 900 Flow vs 707 Flux wings, aftermarket race props, and why everything is still basically stock… for now - Why full-face helmets and downhill MTB armor are becoming mandatory at 33–35 mph - Mental warfare on the beach, prop-wash tactics, hot launches, and pulling 3+G turns - Traveling the world with boards but no batteries (and how the Flightboard rental network saves the day) - The massive progression from the first dealer races in 2022 to riders now training full-time and closing the gap second by second - Where e-foil racing is headed: open-ocean courses, city canal sprints, Everglades gator-chasing, and boards that will eventually hit 50 mph Year one of the Surf Foil World Tour is in the books, prize money is real, brands are paying attention, and the level is skyrocketing. The champ gives us the unfiltered look at what it took to stay on top — and why 2026 is about to get even crazier.If you've ever wondered what the cutting edge of foiling actually looks, sounds, and feels like… this is it.Follow the Surf Foil Tour → https://www.surffoilworldtour.com Justin Chait → https://www.instagram.com/_justinchait_/
Improve your foiling skills in paradise! Join us in Montanita Ecuador May 23-30, 2026 for a foil drive / tow / prone foil camp with Ecuador Foil, KT Foiling & Julia Castro. Learn MoreIn this episode, we sit down with the newly crowned 2025 E-Foil Surf Foil World Tour (SFT) World Champion – the undefeated e-foil racer who took the title in the season finale in Abu Dhabi.Fresh off dominating the inaugural SFT season, the Florida-based ripper (and Flightboard early adopter) joins us to break down what it actually feels like to turn a five-year hunch into a world championship, how e-foil racing went from “nice idea” to a full-blown global tour in record time, and why this sport is exploding faster than anyone predicted.We go deep on:- From kite-smash accidents to building one of the first e-foil schools in South Florida - The wild Atlanta Foil Fest Enduro with Brian Grubb, Nick Leeson, and 20 riders dodging submerged trees at full throttle - Unsanctioned full-send dawn patrols through Amsterdam's canals (don't try this at home) - Gear geek-out: custom shims, chopped tails, 900 Flow vs 707 Flux wings, aftermarket race props, and why everything is still basically stock… for now - Why full-face helmets and downhill MTB armor are becoming mandatory at 33–35 mph - Mental warfare on the beach, prop-wash tactics, hot launches, and pulling 3+G turns - Traveling the world with boards but no batteries (and how the Flightboard rental network saves the day) - The massive progression from the first dealer races in 2022 to riders now training full-time and closing the gap second by second - Where e-foil racing is headed: open-ocean courses, city canal sprints, Everglades gator-chasing, and boards that will eventually hit 50 mph Year one of the Surf Foil World Tour is in the books, prize money is real, brands are paying attention, and the level is skyrocketing. The champ gives us the unfiltered look at what it took to stay on top — and why 2026 is about to get even crazier.If you've ever wondered what the cutting edge of foiling actually looks, sounds, and feels like… this is it.Follow the Surf Foil Tour → https://www.surffoilworldtour.com Justin Chait → https://www.instagram.com/_justinchait_/
This episode is brought to you by Villa Carina Apartments in beautiful Bonaire. In this episode, we catch up with Tom Hartmann – tour manager of the GKA Kite World Tour, Wingfoil World Tour, and founder of the brand-new Surf Foil World Tour (SFT) – fresh from the biggest water sports spectacle the Middle East has ever seen in Abu Dhabi and now chasing the final Kite World Tour stops in Brazil.Tom takes us behind the scenes of the massive nine-day Abu Dhabi event on the soon-to-be “Miami Beach of the Gulf” (Fahid Island), where kite big air, wingfoil racing, e-foil, and wakefoil all shared the spotlight, 2,500+ spectators showed up on weekends, and the whole thing was broadcast live on TV across the region. With €10,000 prize money per SFT discipline, perfect glassy morning conditions, and a level of organization that left athletes speechless, this was the perfect season finale for the inaugural Surf Foil World Tour.- Abu Dhabi deep dive – why foiling (e-foil, wakefoil, wing, kite, surf, pump) is exploding in the Gulf and how the event showcased every flavor of the sport.- E-foil racing at the highest level yet – Justin Chait remains undefeated in 2025, Agnes takes the women's division, and we talk 3G corners, wingtip-out carving, and why technical skill still beats raw speed.- Wakefoil's breakout moment – first fully independent SFT wakefoil comp, drone + boat broadcasting magic, and why wakefoiling could be the most spectator-friendly foiling discipline out there.- The massive growth nobody saw coming – from a hopeful start to nine events worldwide in year one, with a 2026 calendar dropping in the next couple weeks.- What's next for SFT in 2026 – more surf foil, downwind, wakefoil, the return of the epic indoor Düsseldorf pump & wing event, and a brand-new Foil Assist discipline that mixes propulsion take-offs with pure pumping sections.- Plus Tom's love for açaí bowls at Brazilian sunset and maybe sneaking in some surf trips to Nicaragua or Costa Rica before heading home.Year one of the Surf Foil World Tour is officially in the books and it's safe to say foiling just went global – big budgets, big crowds, and bigger stoke. Here's to 2026 being even wilder.Follow the Surf Foil Tour → https://www.surffoiltour.com
Improve your foiling skills in paradise! Join us in Montanita Ecuador May 23-30, 2026 for a foil drive / tow / prone foil camp with Ecuador Foil, KT Foiling & Julia Castro. Learn MoreIn this episode, we catch up with Tom Hartmann – tour manager of the GKA Kite World Tour, Wingfoil World Tour, and founder of the brand-new Surf Foil World Tour (SFT) – fresh from the biggest water sports spectacle the Middle East has ever seen in Abu Dhabi and now chasing the final Kite World Tour stops in Brazil.Tom takes us behind the scenes of the massive nine-day Abu Dhabi event on the soon-to-be “Miami Beach of the Gulf” (Fahid Island), where kite big air, wingfoil racing, e-foil, and wakefoil all shared the spotlight, 2,500+ spectators showed up on weekends, and the whole thing was broadcast live on TV across the region. With €10,000 prize money per SFT discipline, perfect glassy morning conditions, and a level of organization that left athletes speechless, this was the perfect season finale for the inaugural Surf Foil World Tour.- Abu Dhabi deep dive – why foiling (e-foil, wakefoil, wing, kite, surf, pump) is exploding in the Gulf and how the event showcased every flavor of the sport.- E-foil racing at the highest level yet – Justin Chait remains undefeated in 2025, Agnes takes the women's division, and we talk 3G corners, wingtip-out carving, and why technical skill still beats raw speed.- Wakefoil's breakout moment – first fully independent SFT wakefoil comp, drone + boat broadcasting magic, and why wakefoiling could be the most spectator-friendly foiling discipline out there.- The massive growth nobody saw coming – from a hopeful start to nine events worldwide in year one, with a 2026 calendar dropping in the next couple weeks.- What's next for SFT in 2026 – more surf foil, downwind, wakefoil, the return of the epic indoor Düsseldorf pump & wing event, and a brand-new Foil Assist discipline that mixes propulsion take-offs with pure pumping sections.- Plus Tom's love for açaí bowls at Brazilian sunset and maybe sneaking in some surf trips to Nicaragua or Costa Rica before heading home.Year one of the Surf Foil World Tour is officially in the books and it's safe to say foiling just went global – big budgets, big crowds, and bigger stoke. Here's to 2026 being even wilder.Follow the Surf Foil Tour → https://www.surffoiltour.com
In this special release episode, Matt sits down with Nathan Lambert and Luca Soldaini from Ai2 (the Allen Institute for AI) to break down one of the biggest open-source AI drops of the year: OLMo 3. At a moment when most labs are offering “open weights” and calling it a day, AI2 is doing the opposite — publishing the models, the data, the recipes, and every intermediate checkpoint that shows how the system was built. It's an unusually transparent look into the inner machinery of a modern frontier-class model.Nathan and Luca walk us through the full pipeline — from pre-training and mid-training to long-context extension, SFT, preference tuning, and RLVR. They also explain what a thinking model actually is, why reasoning models have exploded in 2025, and how distillation from DeepSeek and Qwen reasoning models works in practice. If you've been trying to truly understand the “RL + reasoning” era of LLMs, this is the clearest explanation you'll hear.We widen the lens to the global picture: why Meta's retreat from open source created a “vacuum of influence,” how Chinese labs like Qwen, DeepSeek, Kimi, and Moonshot surged into that gap, and why so many U.S. companies are quietly building on Chinese open models today. Nathan and Luca offer a grounded, insider view of whether America can mount an effective open-source response — and what that response needs to look like.Finally, we talk about where AI is actually heading. Not the hype, not the doom — but the messy engineering reality behind modern model training, the complexity tax that slows progress, and why the transformation between now and 2030 may be dramatic without ever delivering a single “AGI moment.” If you care about the future of open models and the global AI landscape, this is an essential conversation.Allen Institute for AI (AI2)Website - https://allenai.orgX/Twitter - https://x.com/allen_aiNathan LambertBlog - https://www.interconnects.aiLinkedIn - https://www.linkedin.com/in/natolambert/X/Twitter - https://x.com/natolambertLuca SoldainiBlog - https://soldaini.netLinkedIn - https://www.linkedin.com/in/soldni/X/Twitter - https://x.com/soldniFIRSTMARKWebsite - https://firstmark.comX/Twitter - https://twitter.com/FirstMarkCapMatt Turck (Managing Director)Blog - https://mattturck.comLinkedIn - https://www.linkedin.com/in/turck/X/Twitter - https://twitter.com/mattturck(00:00) – Cold Open(00:39) – Welcome & today's big announcement(01:18) – Introducing the Olmo 3 model family(02:07) – What “base models” really are (and why they matter)(05:51) – Dolma 3: the data behind Olmo 3(08:06) – Performance vs Qwen, Gemma, DeepSeek(10:28) – What true open source means (and why it's rare)(12:51) – Intermediate checkpoints, transparency, and why AI2 publishes everything(16:37) – Why Qwen is everywhere (including U.S. startups)(18:31) – Why Chinese labs go open source (and why U.S. labs don't)(20:28) – Inside ATOM: the U.S. response to China's model surge(22:13) – The rise of “thinking models” and inference-time scaling(35:58) – The full Olmo pipeline, explained simply(46:52) – Pre-training: data, scale, and avoiding catastrophic spikes(50:27) – Mid-training (tail patching) and avoiding test leakage(52:06) – Why long-context training matters(55:28) – SFT: building the foundation for reasoning(1:04:53) – Preference tuning & why DPO still works(1:10:51) – The hard part: RLVR, long reasoning chains, and infrastructure pain(1:13:59) – Why RL is so technically brutal(1:18:17) – Complexity tax vs AGI hype(1:21:58) – How everyone can contribute to the future of AI(1:27:26) – Closing thoughts
Sponsored by Berries: Use code TherapyShow50 for $50 off your first month - CLICK HERE. If you are a therapist or counselor looking for continuing education, check out my NBCC Approved $5 Podcourses and other continuing education offerings. Plus, get your first Podcourse half off. In this episode of The Therapy Show, I'm thrilled to chat with Dr. Tobin Richardson, a fellow continuing education creator and the founder of Save the Therapist. We dive into how Tobin combined his passion for narrative podcasting and his background in counselor education to create a unique, story-driven CE platform for therapists. He shares the behind-the-scenes of launching CE courses that actually engage and inspire, why he believes continuing education needs a serious upgrade, and how he's offering these courses completely free thanks to his creative business model. We also get real about the challenges of building something meaningful, how to market CE content without burning out, and what it's like to cover nuanced, sometimes controversial topics with honesty and integrity. If you've ever thought about creating your own CE content or just want to hear how another therapist is innovating in the field, this episode is for you. Tune in and get inspired by Tobin's journey and maybe even find your next favorite CE podcast. Tobin Richardson, EdD, NCC, is a counselor educator with a decade of experience building and delivering innovative educational resources to therapists in both community mental health and large VC-backed provider organizations. Since launching in early 2025, his NPR-style CE platform Save the Therapist has garnered over 4,000 therapists registrants with over 7,000 course completions. And don't forget! If you're ready to spend less time on notes and more time doing what you love, check out heyberries.com. Use code THERAPYSHOW50 for $50 off your first month with Berries. Get my Coping with Political Stress Ebook and Peaceful Politics AI Guide Therapist Conversation Framework: Politics in Session A printable PDF with 97 questions to navigate political talk in therapy - without taking sides. Solution-Focused Therapy Guide72 questions + prompts to help adult clients clarify goals and move forward using SFT. Check out all my Counselor Resources.
Sponsored by Berries: Use code TherapyShow50 for $50 off your first month - CLICK HERE. If you are a therapist or counselor looking for continuing education, check out my NBCC Approved $5 Podcourses and other continuing education offerings. Plus, get your first Podcourse half off. In this episode, I'm sharing my recent conversation from Between Sessions with Berries, a podcast created for mental health professionals who want to simplify documentation, fight burnout, and reconnect with their purpose. Kym Tolson and I dive into my journey from therapist to continuing-education creator, how burnout inspired me to reimagine CE for busy clinicians, and what it takes to blend creativity, courage, and aligned action into your career. We also discuss lessons learned from losing a major podcast sponsor, building confidence through reinvention, and the power of staying curious in the face of change. If you've ever felt stuck, uninspired, or ready for something new in your professional life, this conversation will encourage you to take the next step - one aligned action at a time. And don't forget! If you're ready to spend less time on notes and more time doing what you love, check out heyberries.com. Use code THERAPYSHOW50 for $50 off your first month with Berries. Get my Coping with Political Stress Ebook and Peaceful Politics AI Guide Therapist Conversation Framework: Politics in Session A printable PDF with 97 questions to navigate political talk in therapy - without taking sides. Solution-Focused Therapy Guide72 questions + prompts to help adult clients clarify goals and move forward using SFT. Check out all my Counselor Resources.
Alex Westphal, Senior Director, Market Practice and Regulatory Policy, talks about the latest milestones in Europe's journey to T+1, also looking at SFT specific impacts and discussions which are central to the success of T+1.
This episode is brought to you by Villa Carina Apartments in beautiful Bonaire. In this episode, we sit down with Tom Hartmann and Nico Hopp of Hoppline to dive into the exhilarating world of pump foiling at Lake Traunsee, Upper Austria. Broadcasting from their respective homes, Tom and Nico share their passion for this rapidly growing sport, the vibrant community, and the unique vibe of the SFT.- Lake Traunsee Triumph: Tom and Nico recap the SFT event at Lake Traunsee, a stunning venue surrounded by mountains with a top-notch setup. With four starting docks and a professional organization running alongside the Austrian Wing Foil Championships, the event offered a perfect mix of competition and community, capped off with exciting wake foiling sessions behind a boat.- Pump Foiling's Appeal: Tom and Nico discuss the sport's accessibility, thriving in flatwater lakes and ideal for urban and inland locations. They highlight how pump foiling draws in everyone from pros to beginners. - Community-Driven Competition: Nico emphasizes the inclusive nature of the SFT, where pros like Eden Fiander and Robert von Roll race alongside amateurs, creating a social and competitive atmosphere. Tom explains the division structure—pro, open, masters, youth, and women's categories—ensuring everyone, from seasoned athletes to first-timers, feels motivated to join. - Gear and Technique Evolution: The duo dives into the latest gear trends, with Nico noting the pros' use of tiny, high-performance wings and unique dock-start techniques. From Eden's strap-based approach to Rob's hands-on style, the diversity in equipment and skills keeps the sport dynamic and exciting. - A Family Affair: Tom highlights the family-friendly vibe, with free dinners for competitors and their families, fostering a welcoming environment. Nico shares a heartwarming story of a young competitor and his mother camping out to participate, showcasing the sport's appeal across generations.- The Future of SFT: Tom reveals plans for the final 2025 event in Abu Dhabi, featuring e-foiling and wake foiling, and a 2026 season kicking off in Düsseldorf. With ambitions to expand prize money and bring events to urban centers like Venice's Grand Canal, the SFT aims to grow pump foiling's global reach.Join us for a lively discussion packed with insights into pump foiling's rise, the thrill of close-knit competition, and the community spirit driving this niche sport forward. From stunning venues to innovative gear, this episode captures the excitement of foiling without wind.Visit: https://www.instagram.com/supfoiltour & https://www.instagram.com/hoppline/
Improve your foiling skills in paradise! Join us in Montanita Ecuador May 23-30, 2026 for a foil drive / tow / prone foil camp with Ecuador Foil, KT Foiling & Julia Castro. Learn MoreIn this episode, we sit down with Tom Hartmann and Nico Hopp of Hoppline to dive into the exhilarating world of pump foiling at Lake Traunsee, Upper Austria. Broadcasting from their respective homes, Tom and Nico share their passion for this rapidly growing sport, the vibrant community, and the unique vibe of the SFT.- Lake Traunsee Triumph: Tom and Nico recap the SFT event at Lake Traunsee, a stunning venue surrounded by mountains with a top-notch setup. With four starting docks and a professional organization running alongside the Austrian Wing Foil Championships, the event offered a perfect mix of competition and community, capped off with exciting wake foiling sessions behind a boat.- Pump Foiling's Appeal: Tom and Nico discuss the sport's accessibility, thriving in flatwater lakes and ideal for urban and inland locations. They highlight how pump foiling draws in everyone from pros to beginners. - Community-Driven Competition: Nico emphasizes the inclusive nature of the SFT, where pros like Eden Fiander and Robert von Roll race alongside amateurs, creating a social and competitive atmosphere. Tom explains the division structure—pro, open, masters, youth, and women's categories—ensuring everyone, from seasoned athletes to first-timers, feels motivated to join. - Gear and Technique Evolution: The duo dives into the latest gear trends, with Nico noting the pros' use of tiny, high-performance wings and unique dock-start techniques. From Eden's strap-based approach to Rob's hands-on style, the diversity in equipment and skills keeps the sport dynamic and exciting. - A Family Affair: Tom highlights the family-friendly vibe, with free dinners for competitors and their families, fostering a welcoming environment. Nico shares a heartwarming story of a young competitor and his mother camping out to participate, showcasing the sport's appeal across generations.- The Future of SFT: Tom reveals plans for the final 2025 event in Abu Dhabi, featuring e-foiling and wake foiling, and a 2026 season kicking off in Düsseldorf. With ambitions to expand prize money and bring events to urban centers like Venice's Grand Canal, the SFT aims to grow pump foiling's global reach.Join us for a lively discussion packed with insights into pump foiling's rise, the thrill of close-knit competition, and the community spirit driving this niche sport forward. From stunning venues to innovative gear, this episode captures the excitement of foiling without wind.Visit: https://www.instagram.com/supfoiltour & https://www.instagram.com/hoppline/
We're pleased to welcome the seasoned driver, Geoffrey 'JB' Njoroge, to the show for this episode of SFT! Today's conversation features a lot of practical wisdom regarding the various opportunities in trucking as well as the daily execution of the craft. JB also recounts some tense and amusing moments from his life behind the wheel as a driver and trainer. You'll enjoy the wit and wisdom of this guest as we once again dive deep into the world of driveaway. If you're not already subscribed to the show, please do so in order to see and hear our weekly content and so you don't miss guys like JB. In fact, he's going to rejoin the show next week to talk about his home country and foundational passions.Show Notes:John and “JB” share a laugh about the exercise challenges over the road (0:44)JB navigates COVID and other experiences in his trucking journey (4:00)Finding Norton and getting “spoiled” in Driveaway (6:08)Crazy stories and sage advice from the Road (9:57)The importance of planning, logs, and dispatch relations (16:12)JB's stats, certifications, and future plans (21:15)Keep Trucking, JB!. The Six-Figure Trucker is a weekly podcast about driveaway trucking brought to you by Norton Transport. For more information or to subscribe, please visit Six-FigureTrucker.com. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
At this year's Regen Gathering on the island of Jersey, our CEO, Patrick Holden, had the chance to meet with the event's co-founder, India Hamilton, for the latest episode of the SFT Podcast. Alongside founding Jersey's Regen Gathering – an annual event which brings together a diverse range of people and ideas to discuss the innovative food, farming and finance approaches that are taking place on Jersey – India is also a chef, food systems expert and heads up HYPHA Consulting, a regenerative consultancy committed to pioneering sustainable futures within the rural economy and food system. In 2018, India was also involved in developing The Sustainable Cooperative (SCOOP), a consumer-led cooperative which aims to create a more sustainable supply of food on Jersey. In this episode, Patrick and India talk about the beginnings of the Jersey Regen Gathering and how its conception was inspired by other food and farming events like Groundswell, what the Jersey government is doing to support their farmers and how this differs from what's happening in the UK, and the connection between public health and our food systems. To connect with India, follow her on LinkedIn. To find out more about the Regen Gathering, visit the website where you can also find details of this year's Jersey Farming Conference, taking place in November. To listen to more SFT podcasts, featuring some of the biggest names in regenerative food and farming, head to our main podcast page. And to keep up to date with our news, you can subscribe to our monthly newsletter or follow us on Instagram, Facebook, LinkedIn and Bluesky. This conversation was recorded in September 2025.
If you are a therapist or counselor looking for continuing education, check out my NBCC Approved $5 Podcourses and other continuing education offerings. Plus, get your first Podcourse half off.
If you are a therapist or counselor looking for continuing education, check out my NBCC Approved $5 Podcourses and other continuing education offerings. Plus, get your first Podcourse half off. Check out all my Counselor Resources.
In this episode we catch up with Tom Hartmann about the Atlanta Foil Fest at Lake Lanier's Olympic Park. Tom dives into the three-day event, featuring E-Foil, Pump Foil, Wake Foil, and Airchair foiling competitions, alongside demos and clinics. From Justin Chait dominant performance to Nick Leason's E-Foil legacy and the innovative Betafoil Enduro, the event united foilers from the US, Europe, and beyond. Tom also teases upcoming Surf World Tour events at Lake Garda and Abu Dhabi, promising bigger competitions and live streams.Episode Highlights:Atlanta Foil Fest's debut as a foiling mecca with a custom-built start dockHigh-level E-Foil and Pump Foil races, plus the return of AirchairMeeting foiling pioneer Nick Leeson and testing Betafoil's massive wingsLake Garda's Foiling Week and Abu Dhabi's massive prize purse on the horizonGrowing the global foiling community through passion and connectionFollow Tom & SFT: @surffoilworldtour on Instagram, Facebook, YouTubeWebsite: surffoilworldtour.com
Improve your foiling skills in paradise! Join us in Montanita Ecuador May 23-30, 2026 for a foil drive / tow / prone foil camp with Ecuador Foil, KT Foiling & Julia Castro. Learn MoreIn this episode we catch up with Tom Hartmann about the Atlanta Foil Fest at Lake Lanier's Olympic Park. Tom dives into the three-day event, featuring E-Foil, Pump Foil, Wake Foil, and Airchair foiling competitions, alongside demos and clinics. From Justin Chait dominant performance to Nick Leason's E-Foil legacy and the innovative Betafoil Enduro, the event united foilers from the US, Europe, and beyond. Tom also teases upcoming Surf World Tour events at Lake Garda and Abu Dhabi, promising bigger competitions and live streams.Episode Highlights:Atlanta Foil Fest's debut as a foiling mecca with a custom-built start dockHigh-level E-Foil and Pump Foil races, plus the return of AirchairMeeting foiling pioneer Nick Leeson and testing Betafoil's massive wingsLake Garda's Foiling Week and Abu Dhabi's massive prize purse on the horizonGrowing the global foiling community through passion and connectionFollow Tom & SFT: @surffoilworldtour on Instagram, Facebook, YouTubeWebsite: surffoilworldtour.com
ffinlo Costain (8point9.com) and Joe Stanley (GWCT Allerton Project) discuss:Net zero reports from The Tony Blair Institute and the AFN Network+UK climate change preparednessUK Government 'retakes' the decision to scrap SFI 2024Anaerobic digestionAnd those reports - by FAI and SFT - that were damned by Monbiot.
Improve your foiling skills in paradise! Join us in Montanita Ecuador May 23-30, 2026 for a foil drive / tow / prone foil camp with Ecuador Foil, KT Foiling & Julia Castro. Learn MoreWe're thrilled to kick off our new role as the official podcast of the Surf Foil World Tour (SFT)! In this episode, we sit down with Jørgen Vogt and Tom Hartmann, the masterminds behind the SFT, a groundbreaking world tour for non-wind-powered foiling disciplines. From its roots in the GKA (Kite World Tour) and GWA (Wing Foil World Tour) to embracing pump foiling, surf foiling, and more, Jørgen and Tom share how the SFT is uniting passionate amateurs and pros alike, fostering a vibrant foiling community. This chat dives into the tour's grassroots vibe and big ambitions.In this episode, you'll discover:SFT's Origin Story: How a 2019 Cape Town car chat sparked a new tour, building on Jørgen's GKA success and Tom's windsurf tour experience.Five Core Disciplines: Pump foiling, surf foiling, downwind foiling, e-foiling, and wake foiling—accessible sports for lake lovers, yacht owners, and wave chasers.Grassroots Appeal: Why the SFT's Pro-Am approach welcomes enthusiasts, not just pros, fostering friendships and community at events like Pensacola and Sicily.Global Growth: From Düsseldorf's indoor World Cup to North America's rising demand, with the Atlanta Foil Fast World Cup (June 13-15, 2025) on the horizon.Event Challenges: Navigating financial hurdles and explaining niche foiling to sponsors, unlike the established GKA and GWA tours.Jørgen & Tom's Partnership: A chance meeting at a Mauritius Kite World Cup party led to a dynamic duo driving foiling's future.Community Impact: How SFT events, like Sicily's Water Experience, invite newcomers to try foiling, sparking passion across generations.Check out the SFT at https://www.surffoilworldtour.com/ for event schedules and details. Join the foiling community and catch the Atlanta Foil Fest World Cup vibe!
We're thrilled to kick off our new role as the official podcast of the Surf Foil World Tour (SFT)! In this episode, we sit down with Jørgen Vogt and Tom Hartmann, the masterminds behind the SFT, a groundbreaking world tour for non-wind-powered foiling disciplines. From its roots in the GKA (Kite World Tour) and GWA (Wing Foil World Tour) to embracing pump foiling, surf foiling, and more, Jørgen and Tom share how the SFT is uniting passionate amateurs and pros alike, fostering a vibrant foiling community. This chat dives into the tour's grassroots vibe and big ambitions.In this episode, you'll discover:SFT's Origin Story: How a 2019 Cape Town car chat sparked a new tour, building on Jørgen's GKA success and Tom's windsurf tour experience.Five Core Disciplines: Pump foiling, surf foiling, downwind foiling, e-foiling, and wake foiling—accessible sports for lake lovers, yacht owners, and wave chasers.Grassroots Appeal: Why the SFT's Pro-Am approach welcomes enthusiasts, not just pros, fostering friendships and community at events like Pensacola and Sicily.Global Growth: From Düsseldorf's indoor World Cup to North America's rising demand, with the Atlanta Foil Fast World Cup (June 13-15, 2025) on the horizon.Event Challenges: Navigating financial hurdles and explaining niche foiling to sponsors, unlike the established GKA and GWA tours.Jørgen & Tom's Partnership: A chance meeting at a Mauritius Kite World Cup party led to a dynamic duo driving foiling's future.Community Impact: How SFT events, like Sicily's Water Experience, invite newcomers to try foiling, sparking passion across generations.Check out the SFT at https://www.surffoilworldtour.com/ for event schedules and details. Join the foiling community and catch the Atlanta Foil Fest World Cup vibe!
This Week in Machine Learning & Artificial Intelligence (AI) Podcast
Today, we're joined by Mahesh Sathiamoorthy, co-founder and CEO of Bespoke Labs, to discuss how reinforcement learning (RL) is reshaping the way we build custom agents on top of foundation models. Mahesh highlights the crucial role of data curation, evaluation, and error analysis in model performance, and explains why RL offers a more robust alternative to prompting, and how it can improve multi-step tool use capabilities. We also explore the limitations of supervised fine-tuning (SFT) for tool-augmented reasoning tasks, the reward-shaping strategies they've used, and Bespoke Labs' open-source libraries like Curator. We also touch on the models MiniCheck for hallucination detection and MiniChart for chart-based QA. The complete show notes for this episode can be found at https://twimlai.com/go/731.
In this episode, Amit and Dheeraj dive deep into the world of AI reasoning models with Alex, an AI researcher involved in OpenThinker and OpenThoughts. They explore two recent groundbreaking papers—SkyT1 and S1 (Simple Test Time Scaling)—that showcase new insights into how large language models (LLMs) develop reasoning capabilities.From structured reasoning vs. content accuracy to fine-tuning efficiency and the role of active learning, this conversation highlights the shift from prompt engineering to structured supervised fine-tuning (SFT) and post-training techniques. The discussion also touches on open weights, open data, and open-source AI, revealing the evolving AI landscape and its impact on startups, research, and beyond.Key Topics & Chapter Markers[00:00] Introduction – Why reasoning models matter & today's agenda[05:15] Breaking Down SkyT1 – Structure vs. Content in reasoning[15:45] Open weights, open data, and open-source AI[22:30] Fine-tuning vs. RL – When do you need reinforcement learning?[30:10] S1 and the power of test-time scaling[40:25] Budget forcing – Making AI "think" more efficiently[50:50] RAG vs. SFT – What should startups use?[01:05:30] Active learning – AI asking the right questions[01:15:00] Final thoughts – Where AI reasoning is heading nextResources & Links
Seek For Truth 2.3.25 |Ancient Destrruction | Acts of peter | Dying testimoniesVisit us at www.seekfortruth.org
Happy holidays! We'll be sharing snippets from Latent Space LIVE! through the break bringing you the best of 2024! We want to express our deepest appreciation to event sponsors AWS, Daylight Computer, Thoth.ai, StrongCompute, Notable Capital, and most of all all our LS supporters who helped fund the gorgeous venue and A/V production!For NeurIPS last year we did our standard conference podcast coverage interviewing selected papers (that we have now also done for ICLR and ICML), however we felt that we could be doing more to help AI Engineers 1) get more industry-relevant content, and 2) recap 2024 year in review from experts. As a result, we organized the first Latent Space LIVE!, our first in person miniconference, at NeurIPS 2024 in Vancouver. Today, we're proud to share Loubna's highly anticipated talk (slides here)!Synthetic DataWe called out the Synthetic Data debate at last year's NeurIPS, and no surprise that 2024 was dominated by the rise of synthetic data everywhere:* Apple's Rephrasing the Web, Microsoft's Phi 2-4 and Orca/AgentInstruct, Tencent's Billion Persona dataset, DCLM, and HuggingFace's FineWeb-Edu, and Loubna's own Cosmopedia extended the ideas of synthetic textbook and agent generation to improve raw web scrape dataset quality* This year we also talked to the IDEFICS/OBELICS team at HuggingFace who released WebSight this year, the first work on code-vs-images synthetic data.* We called Llama 3.1 the Synthetic Data Model for its extensive use (and documentation!) of synthetic data in its pipeline, as well as its permissive license. * Nemotron CC and Nemotron-4-340B also made a big splash this year for how they used 20k items of human data to synthesize over 98% of the data used for SFT/PFT.* Cohere introduced Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress observing gains of up to 56.5% improvement in win rates comparing multiple teachers vs the single best teacher model* In post training, AI2's Tülu3 (discussed by Luca in our Open Models talk) and Loubna's Smol Talk were also notable open releases this year.This comes in the face of a lot of scrutiny and criticism, with Scale AI as one of the leading voices publishing AI models collapse when trained on recursively generated data in Nature magazine bringing mainstream concerns to the potential downsides of poor quality syndata:Part of the concerns we highlighted last year on low-background tokens are coming to bear: ChatGPT contaminated data is spiking in every possible metric:But perhaps, if Sakana's AI Scientist pans out this year, we will have mostly-AI AI researchers publishing AI research anyway so do we really care as long as the ideas can be verified to be correct?Smol ModelsMeta surprised many folks this year by not just aggressively updating Llama 3 and adding multimodality, but also adding a new series of “small” 1B and 3B “on device” models this year, even working on quantized numerics collaborations with Qualcomm, Mediatek, and Arm. It is near unbelievable that a 1B model today can qualitatively match a 13B model of last year:and the minimum size to hit a given MMLU bar has come down roughly 10x in the last year. We have been tracking this proxied by Lmsys Elo and inference price:The key reads this year are:* MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases* Apple Intelligence Foundation Language Models* Hymba: A Hybrid-head Architecture for Small Language Models* Loubna's SmolLM and SmolLM2: a family of state-of-the-art small models with 135M, 360M, and 1.7B parameters on the pareto efficiency frontier.* and Moondream, which we already covered in the 2024 in Vision talkFull Talk on YouTubeplease like and subscribe!Timestamps* [00:00:05] Loubna Intro* [00:00:33] The Rise of Synthetic Data Everywhere* [00:02:57] Model Collapse* [00:05:14] Phi, FineWeb, Cosmopedia - Synthetic Textbooks* [00:12:36] DCLM, Nemotron-CC* [00:13:28] Post Training - AI2 Tulu, Smol Talk, Cohere Multilingual Arbitrage* [00:16:17] Smol Models* [00:18:24] On Device Models* [00:22:45] Smol Vision Models* [00:25:14] What's NextTranscript2024 in Synthetic Data and Smol Models[00:00:00] [00:00:05] Loubna Intro[00:00:05] Speaker: I'm very happy to be here. Thank you for the invitation. So I'm going to be talking about synthetic data in 2024. And then I'm going to be talking about small on device models. So I think the most interesting thing about synthetic data this year is that like now we have it everywhere in the large language models pipeline.[00:00:33] The Rise of Synthetic Data Everywhere[00:00:33] Speaker: I think initially, synthetic data was mainly used just for post training, because naturally that's the part where we needed human annotators. And then after that, we realized that we don't really have good benchmarks to [00:01:00] measure if models follow instructions well, if they are creative enough, or if they are chatty enough, so we also started using LLMs as judges.[00:01:08] Speaker: Thank you. And I think this year and towards the end of last year, we also went to the pre training parts and we started generating synthetic data for pre training to kind of replace some parts of the web. And the motivation behind that is that you have a lot of control over synthetic data. You can control your prompt and basically also the kind of data that you generate.[00:01:28] Speaker: So instead of just trying to filter the web, you could try to get the LLM to generate what you think the best web pages could look like and then train your models on that. So this is how we went from not having synthetic data at all in the LLM pipeline to having it everywhere. And so the cool thing is like today you can train an LLM with like an entirely synthetic pipeline.[00:01:49] Speaker: For example, you can use our Cosmopedia datasets and you can train a 1B model on like 150 billion tokens that are 100 percent synthetic. And those are also of good quality. And then you can [00:02:00] instruction tune the model on a synthetic SFT dataset. You can also do DPO on a synthetic dataset. And then to evaluate if the model is good, you can use.[00:02:07] Speaker: A benchmark that uses LLMs as a judge, for example, MTBench or AlpacaEvil. So I think this is like a really mind blowing because like just a few years ago, we wouldn't think this is possible. And I think there's a lot of concerns about model collapse, and I'm going to talk about that later. But we'll see that like, if we use synthetic data properly and we curate it carefully, that shouldn't happen.[00:02:29] Speaker: And the reason synthetic data is very popular right now is that we have really strong models, both open and closed. It is really cheap and fast to use compared to human annotations, which cost a lot and take a lot of time. And also for open models right now, we have some really good inference frameworks.[00:02:47] Speaker: So if you have enough GPUs, it's really easy to spawn these GPUs and generate like a lot of synthetic data. Some examples are VLM, TGI, and TensorRT.[00:02:57] Model Collapse[00:02:57] Speaker: Now let's talk about the elephant in the room, model [00:03:00] collapse. Is this the end? If you look at the media and all of like, for example, some papers in nature, it's really scary because there's a lot of synthetic data out there in the web.[00:03:09] Speaker: And naturally we train on the web. So we're going to be training a lot of synthetic data. And if model collapse is going to happen, we should really try to take that seriously. And the other issue is that, as I said, we think, a lot of people think the web is polluted because there's a lot of synthetic data.[00:03:24] Speaker: And for example, when we're building fine web datasets here at Guillerm and Hinek, we're interested in like, how much synthetic data is there in the web? So there isn't really a method to properly measure the amount of synthetic data or to save a webpage synthetic or not. But one thing we can do is to try to look for like proxy words, for example, expressions like as a large language model or words like delve that we know are actually generated by chat GPT.[00:03:49] Speaker: We could try to measure the amount of these words in our data system and compare them to the previous years. For example, here, we measured like a, these words ratio in different dumps of common crawl. [00:04:00] And we can see that like the ratio really increased after chat GPT's release. So if we were to say that synthetic data amount didn't change, you would expect this ratio to stay constant, which is not the case.[00:04:11] Speaker: So there's a lot of synthetic data probably on the web, but does this really make models worse? So what we did is we trained different models on these different dumps. And we then computed their performance on popular, like, NLP benchmarks, and then we computed the aggregated score. And surprisingly, you can see that the latest DOMs are actually even better than the DOMs that are before.[00:04:31] Speaker: So if there's some synthetic data there, at least it did not make the model's worse. Yeah, which is really encouraging. So personally, I wouldn't say the web is positive with Synthetic Data. Maybe it's even making it more rich. And the issue with like model collapse is that, for example, those studies, they were done at like a small scale, and you would ask the model to complete, for example, a Wikipedia paragraph, and then you would train it on these new generations, and you would do that every day.[00:04:56] Speaker: iteratively. I think if you do that approach, it's normal to [00:05:00] observe this kind of behavior because the quality is going to be worse because the model is already small. And then if you train it just on its generations, you shouldn't expect it to become better. But what we're really doing here is that we take a model that is very large and we try to distill its knowledge into a model that is smaller.[00:05:14] Phi, FineWeb, Cosmopedia - Synthetic Textbooks[00:05:14] Speaker: And in this way, you can expect to get like a better performance for your small model. And using synthetic data for pre-training has become really popular. After the textbooks are all you need papers where Microsoft basically trained a series of small models on textbooks that were using a large LLM.[00:05:32] Speaker: And then they found that these models were actually better than models that are much larger. So this was really interesting. It was like first of its time, but it was also met with a lot of skepticism, which is a good thing in research. It pushes you to question things because the dataset that they trained on was not public, so people were not really sure if these models are really good or maybe there's just some data contamination.[00:05:55] Speaker: So it was really hard to check if you just have the weights of the models. [00:06:00] And as Hugging Face, because we like open source, we tried to reproduce what they did. So this is our Cosmopedia dataset. We basically tried to follow a similar approach to what they documented in the paper. And we created a synthetic dataset of textbooks and blog posts and stories that had almost 30 billion tokens.[00:06:16] Speaker: And we tried to train some models on that. And we found that like the key ingredient to getting a good data set that is synthetic is trying as much as possible to keep it diverse. Because if you just throw the same prompts as your model, like generate like a textbook about linear algebra, and even if you change the temperature, the textbooks are going to look alike.[00:06:35] Speaker: So there's no way you could scale to like millions of samples. And the way you do that is by creating prompts that have some seeds that make them diverse. In our case, the prompt, we would ask the model to generate a textbook, but make it related to an extract from a webpage. And also we try to frame it within, to stay within topic.[00:06:55] Speaker: For example, here, we put like an extract about cardiovascular bioimaging, [00:07:00] and then we ask the model to generate a textbook related to medicine that is also related to this webpage. And this is a really nice approach because there's so many webpages out there. So you can. Be sure that your generation is not going to be diverse when you change the seed example.[00:07:16] Speaker: One thing that's challenging with this is that you want the seed samples to be related to your topics. So we use like a search tool to try to go all of fine web datasets. And then we also do a lot of experiments with the type of generations we want the model to generate. For example, we ask it for textbooks for middle school students or textbook for college.[00:07:40] Speaker: And we found that like some generation styles help on some specific benchmarks, while others help on other benchmarks. For example, college textbooks are really good for MMLU, while middle school textbooks are good for benchmarks like OpenBookQA and Pico. This is like a sample from like our search tool.[00:07:56] Speaker: For example, you have a top category, which is a topic, and then you have some [00:08:00] subtopics, and then you have the topic hits, which are basically the web pages in fine web does belong to these topics. And here you can see the comparison between Cosmopedia. We had two versions V1 and V2 in blue and red, and you can see the comparison to fine web, and as you can see throughout the training training on Cosmopedia was consistently better.[00:08:20] Speaker: So we managed to get a data set that was actually good to train these models on. It's of course so much smaller than FineWeb, it's only 30 billion tokens, but that's the scale that Microsoft data sets was, so we kind of managed to reproduce a bit what they did. And the data set is public, so everyone can go there, check if everything is all right.[00:08:38] Speaker: And now this is a recent paper from NVIDIA, Neumatron CC. They took things a bit further, and they generated not a few billion tokens, but 1. 9 trillion tokens, which is huge. And we can see later how they did that. It's more of, like, rephrasing the web. So we can see today that there's, like, some really huge synthetic datasets out there, and they're public, so, [00:09:00] like, you can try to filter them even further if you want to get, like, more high quality corpses.[00:09:04] Speaker: So for this, rephrasing the web this approach was suggested in this paper by Pratyush, where basically in this paper, they take some samples from C4 datasets, and then they use an LLM to rewrite these samples into a better format. For example, they ask an LLM to rewrite the sample into a Wikipedia passage or into a Q& A page.[00:09:25] Speaker: And the interesting thing in this approach is that you can use a model that is Small because it doesn't, rewriting doesn't require knowledge. It's just rewriting a page into a different style. So the model doesn't need to have like knowledge that is like extensive of what is rewriting compared to just asking a model to generate a new textbook and not giving it like ground truth.[00:09:45] Speaker: So here they rewrite some samples from C4 into Q& A, into Wikipedia, and they find that doing this works better than training just on C4. And so what they did in Nemo Trans CC is a similar approach. [00:10:00] They rewrite some pages from Common Crawl for two reasons. One is to, like improve Pages that are low quality, so they rewrite them into, for example, Wikipedia page, so they look better.[00:10:11] Speaker: And another reason is to create more diverse datasets. So they have a dataset that they already heavily filtered, and then they take these pages that are already high quality, and they ask the model to rewrite them in Question and Answer format. into like open ended questions or like multi choice questions.[00:10:27] Speaker: So this way they can reuse the same page multiple times without fearing like having multiple duplicates, because it's the same information, but it's going to be written differently. So I think that's also a really interesting approach for like generating synthetic data just by rephrasing the pages that you already have.[00:10:44] Speaker: There's also this approach called Prox where they try to start from a web page and then they generate a program which finds how to write that page to make it better and less noisy. For example, here you can see that there's some leftover metadata in the web page and you don't necessarily want to keep that for training [00:11:00] your model.[00:11:00] Speaker: So So they train a model that can generate programs that can like normalize and remove lines that are extra. So I think this approach is also interesting, but it's maybe less scalable than the approaches that I presented before. So that was it for like rephrasing and generating new textbooks.[00:11:17] Speaker: Another approach that I think is really good and becoming really popular for using synthetic data for pre training is basically building a better classifiers. For filtering the web for example, here we release the data sets called fine web edu. And the way we built it is by taking Llama3 and asking it to rate the educational content of web pages from zero to five.[00:11:39] Speaker: So for example, if a page is like a really good textbook that could be useful in a school setting, it would get a really high score. And if a page is just like an advertisement or promotional material, it would get a lower score. And then after that, we take these synthetic annotations and we train a classifier on them.[00:11:57] Speaker: It's a classifier like a BERT model. [00:12:00] And then we run this classifier on all of FineWeb, which is a 15 trillion tokens dataset. And then we only keep the pages that have like a score that's higher than 3. So for example, in our case, we went from 15 trillion tokens to 3. to just 1. 5 trillion tokens. Those are really highly educational.[00:12:16] Speaker: And as you can see here, a fine web EDU outperforms all the other public web datasets by a larger margin on a couple of benchmarks here, I show the aggregated score and you can see that this approach is really effective for filtering web datasets to get like better corpuses for training your LLMs.[00:12:36] DCLM, Nemotron-CC[00:12:36] Speaker: Others also try to do this approach. There's, for example, the DCLM datasets where they also train the classifier, but not to detect educational content. Instead, they trained it on OpenHermes dataset, which is a dataset for instruction tuning. And also they explain like IAM5 subreddits, and then they also get really high quality dataset which is like very information dense and can help [00:13:00] you train some really good LLMs.[00:13:01] Speaker: And then Nemotron Common Crawl, they also did this approach, but instead of using one classifier, they used an ensemble of classifiers. So they used, for example, the DCLM classifier, and also classifiers like the ones we used in FineWebEducational, and then they combined these two. Scores into a, with an ensemble method to only retain the best high quality pages, and they get a data set that works even better than the ones we develop.[00:13:25] Speaker: So that was it for like synthetic data for pre-training.[00:13:28] Post Training - AI2 Tulu, Smol Talk, Cohere Multilingual Arbitrage[00:13:28] Speaker: Now we can go back to post training. I think there's a lot of interesting post training data sets out there. One that was released recently, the agent instructs by Microsoft where they basically try to target some specific skills. And improve the performance of models on them.[00:13:43] Speaker: For example, here, you can see code, brain teasers, open domain QA, and they managed to get a dataset that outperforms that's when fine tuning Mistral 7b on it, it outperforms the original instruct model that was released by Mistral. And as I said, to get good synthetic data, you really [00:14:00] have to have a framework to make sure that your data is diverse.[00:14:03] Speaker: So for example, for them, they always. And then they see the generations on either source code or raw text documents, and then they rewrite them to make sure they're easier to generate instructions from, and then they use that for their like instruction data generation. There's also the Tool3SFT mixture, which was released recently by Allen AI.[00:14:23] Speaker: It's also really good quality and it covers a wide range of tasks. And the way they make sure that this dataset is diverse is by using personas from the persona hub datasets. Which is basically a data set of like I think over a million personas. And for example, in the tool mixture to generate like a new code snippet, they would give like the model persona, for example, a machine learning researcher interested in neural networks, and then ask it to generate like a coding problem.[00:14:49] Speaker: This way you make sure that your data set is really diverse, and then you can further filter the data sets, for example, using the reward models. We also released a dataset called Smalltalk, [00:15:00] and we also tried to cover the wide range of tasks, and as you can see here, for example, when fine tuning Mistral 7b on the dataset, we also outperformed the original Mistral instructs on a number of benchmarks, notably on mathematics and instruction following with ifevil.[00:15:18] Speaker: Another paper that's really interesting I wanted to mention is this one called Multilingual Data Arbitrage by Cohere. And basically they want to generate a data set for post training that is multilingual. And they have a really interesting problem. It's the fact that there isn't like one model that's really good at all the languages they wanted.[00:15:36] Speaker: So what they do is that like they use not just one teacher model, but multiple teachers. And then they have a router which basically sends the prompts they have to all these models. And then they get the completions and they have a reward model that traces all these generations and only keeps the best one.[00:15:52] Speaker: And this is like arbitrage and finance. So well, I think what's interesting in this, it shows that like synthetic data, it doesn't have to come from a single model. [00:16:00] And because we have so many good models now, you could like pull these models together and get like a dataset that's really high quality and that's diverse and that's covers all your needs.[00:16:12] Speaker: I was supposed to put a meme there, but. Yeah, so that was it for like a synthetic data.[00:16:17] Smol Models[00:16:17] Speaker: Now we can go to see what's happening in the small models field in 2024. I don't know if you know, but like now we have some really good small models. For example, Lama 3. 2 1B is. It matches Lama 2. 13b from, that was released last year on the LMSYS arena, which is basically the default go to leaderboard for evaluating models using human evaluation.[00:16:39] Speaker: And as you can see here, the scores of the models are really close. So I think we've made like hugely forward in terms of small models. Of course, that's one, just one data point, but there's more. For example, if you look at this chart from the Quint 2. 5 blog post, it shows that today we have some really good models that are only like 3 billion parameters [00:17:00] and 4 billion that score really high on MMLU.[00:17:03] Speaker: Which is a really popular benchmark for evaluating models. And you can see here that the red, the blue dots have more than 65 on MMLU. And the grey ones have less. And for example, Llama33b had less. So now we have a 3b model that outperforms a 33b model that was released earlier. So I think now people are starting to realize that like, we shouldn't just scale and scale models, but we should try to make them more efficient.[00:17:33] Speaker: I don't know if you knew, but you can also chat with a 3B plus model on your iPhone. For example, here, this is an app called PocketPal, where you can go and select a model from Hugging Face. It has a large choice. For example, here we loaded the 5. 3. 5, which is 3. 8 billion parameters on this iPhone. And we can chat with this and you can see that even the latency is also acceptable.[00:17:57] Speaker: For example, here, I asked it to give me a joke about [00:18:00] NeurIPS. So let's see what it has to say.[00:18:06] Speaker: Okay, why did the neural network attend NeurIPS? Because it heard there would be a lot of layers and fun and it wanted to train its sense of humor. So not very funny, but at least it can run on device. Yeah, so I think now we have good small models, but we also have like good frameworks and tools to use these small models.[00:18:24] On Device Models[00:18:24] Speaker: So I think we're really close to having like really on edge and on device models that are really good. And I think for a while we've had this narrative. But just training larger models is better. Of course, this is supported by science scaling laws. As you can see here, for example, when we scale the model size, the loss is lower and obviously you get a better model.[00:18:46] Speaker: But and we can see this, for example, in the GPT family of models, how we went from just a hundred million parameters to more than a trillion. parameters. And of course, we all observed the performance improvement when using the latest model. But [00:19:00] one thing that we shouldn't forget is that when we scale the model, we also scale the inference costs and time.[00:19:05] Speaker: And so the largest models were are going to cost so much more. So I think now instead of just building larger models, we should be focusing on building more efficient models. It's no longer a race for the largest models since these models are really expensive to run and they require like a really good infrastructure to do that and they cannot run on, for example, consumer hardware.[00:19:27] Speaker: And when you try to build more efficient models that match larger models, that's when you can really unlock some really interesting on device use cases. And I think a trend that we're noticing now is the trend of training smaller models longer. For example, if you compare how much, how long LLAMA was trained compared to LLAMA3, there is a huge increase in the pre training length.[00:19:50] Speaker: LLAMA was trained on 1 trillion tokens, but LLAMA3 8b was trained on 15 trillion tokens. So Meta managed to get a model that's the same size, but But it performs so much [00:20:00] better by choosing to like spend the sacrifice during training, because as we know, training is a one time cost, but inference is something that's ongoing.[00:20:08] Speaker: If we want to see what are like the small models reads in 2024, I think this mobile LLM paper by Meta is interesting. They try to study different models that are like have the less than 1 billion parameters and find which architecture makes most sense for these models. For example, they find that depth is more important than width.[00:20:29] Speaker: So it's more important to have models that have like more layers than just one. making them more wide. They also find that GQA helps, that tying the embedding helps. So I think it's a nice study overall for models that are just a few hundred million parameters. There's also the Apple intelligence tech report, which is interesting.[00:20:48] Speaker: So for Apple intelligence, they had two models, one that was like on server and another model that was on device. It had 3 billion parameters. And I think the interesting part is that they trained this model using [00:21:00] pruning. And then distillation. And for example, they have this table where they show that, like, using pruning and distillation works much better than training from scratch.[00:21:08] Speaker: And they also have some interesting insights about, like, how they specialize their models on specific tasks, like, for example, summarization and rewriting. There's also this paper by NVIDIA that was released recently. I think you've already had a talk about, like, hybrid models that was all interesting.[00:21:23] Speaker: And this model, they used, like, a hybrid architecture between state space models and transformers. And they managed to train a 1B model that's really performant without needing to train it on a lot of tokens. And regarding our work, we just recently released SmallM2, so it's a series of three models, which are the best in class in each model size.[00:21:46] Speaker: For example, our 1. 7b model outperforms Lama 1b and also Qt 2. 5. And how we managed to train this model is the following. That's where you spent a lot of time trying to curate the pre training datasets. We did a lot of [00:22:00] ablations, trying to find which datasets are good and also how to mix them. We also created some new math and code datasets that we're releasing soon.[00:22:08] Speaker: But you basically really spent a lot of time trying to find what's the best mixture that you can train these models on. And then we spent some time trying to like we also trained these models for very long. For example, small M1 was trained only on 1 trillion tokens, but this model is trained on 11 trillion tokens.[00:22:24] Speaker: And we saw that the performance kept improving. The models didn't really plateau mid training, which I think is really interesting. It shows that you can train such small models for very long and keep getting performance gains. What's interesting about SmallLM2 is that it's fully open. We also released, like the pre training code base, the fine tuning code, the datasets, and also evaluation in this repository.[00:22:45] Smol Vision Models[00:22:45] Speaker: Also there's, like, really interesting small models for text, but also for vision. For example, here you can see SmallVLM, which is a 2B model that's really efficient. It doesn't consume a lot of RAM, and it also has a good performance. There's also Moondream 0. [00:23:00] 5b, which was released recently. It's like the smallest visual language model.[00:23:04] Speaker: And as you can see, there isn't like a big trade off compared to Moondream 2b. So now I showed you that we have some really good small models. We also have the tools to use them, but why should you consider using small models and when? I think, like, small models are really interesting because of the on device feature.[00:23:23] Speaker: Because these models are small and they can run fast, you can basically run them on your laptop, but also on your mobile phone. And this means that your dataset stays locally. You don't have to send your queries to third parties. And this really enhances privacy. That was, for example, one of the big selling points for Apple Intelligence.[00:23:42] Speaker: Also, right now, we really have a lot of work to do. So many frameworks to do on device inference. For example, there's MLX, MLC, Llama, CPP, Transformers, JS. So we have a lot of options and each of them have like great features. So you have so many options for doing that. Small models are also really powerful if you choose to specialize them.[00:24:00][00:24:00] Speaker: For example, here there's a startup called Numind, which took small LM and then they fine tuned it on text extraction datasets. And they managed to get a model that's not very far from models that are much larger. So I think text extraction is like one use case where small models can be really performant and it makes sense to use them instead of just using larger models.[00:24:19] Speaker: You can also chat with these models in browser. For example, here, you can go there, you can load the model, you can even turn off your internet and just start chatting with the model locally. Speaking of text extraction, if you don't want to fine tune the models, there's a really good method of structure generation.[00:24:36] Speaker: We can basically force the models to follow a JSON schema that you defined. For example, here, we try to force the model to follow a schema for extracting key information from GitHub issues. So you can input free text, which is a complaint about a GitHub repository, something not working. And then you can run it there and the model can extract anything that is relevant for your GitHub issue creation.[00:24:58] Speaker: For example, the [00:25:00] priority, for example, here, priority is high, the type of the issue bug, and then a title and the estimation of how long this will take to fix. And you can just like do this in the browser, you can transform your text into a GitHub issue that's properly formatted.[00:25:14] What's Next[00:25:14] Speaker: So what's next for synthetic data and small models?[00:25:18] Speaker: I think that domain specific synthetic data is going to be, it's already important, it's going to be even more important. For example, generating synthetic data for math. I think this really would help improve the reasoning of a lot of models. And a lot of people are doing it, for example, Quint 2. 12 math, everyone's trying to reproduce a one.[00:25:37] Speaker: And so I think for synthetic data, trying to specialize it on some domains is going to be really important. And then for small models, I think specializing them through fine tuning, it's also going to be really important because I think a lot of companies are just trying to use these large models because they are better.[00:25:53] Speaker: But on some tasks, I think you can already get decent performance with small models. So you don't need to Pay like a [00:26:00] cost that's much larger just to make your model better at your task by a few percent. And this is not just for text. And I think it also applies for other modalities like vision and audio.[00:26:11] Speaker: And I think you should also watch out for on device frameworks and applications. For example, like the app I showed, or lama, all these frameworks are becoming really popular and I'm pretty sure that we're gonna get like more of them in 2025. And users really like that. Maybe for other, I should also say hot take.[00:26:28] Speaker: I think that like in AI, we just started like with fine tuning, for example, trying to make BERT work on some specific use cases, and really struggling to do that. And then we had some models that are much larger. So we just switched to like prompt engineering to get the models And I think we're going back to fine tuning where we realize these models are really costly.[00:26:47] Speaker: It's better to use just a small model or try to specialize it. So I think it's a little bit of a cycle and we're going to start to see like more fine tuning and less of just like a prompt engineering the models. So that was my talk. Thank you for following. And if you have [00:27:00] any questions, we can take them now. Get full access to Latent Space at www.latent.space/subscribe
(Weihnachtswoche, Tag 4) "Wusstet ihr eigentlich …?" ist unser Name für "Zusatzfolgen" zu großen Podcasts. In ihnen kommen alle Infos, Anekdoten, O-Töne und andere Dinge unter, die in der Hauptfolge keinen Platz hatten, vergessen wurden oder uns zu spät eingefallen sind. Das Format ist eines unserer beliebesten und häufigsten Unterstützerformate: Allein in diesem Jahr gab es (mit dieser) 16 Ausgaben davon. 6 davon bezogen sich auf SF-Folgen, 7 auf SSF und 3 auf SFT. Diesmal geht es um Deus Ex, das ist keine Wiederveröffentlichung aus dem Steady- oder Patreon-Feed, sondern eine brandneue Folge! Darin besprechen wir… … das Geheimnis des fehlenden Empire State Buildings … warum die Leitern im Spiel so schlecht funktionieren … den Besuch der Matrix und der Ion-Storm-Büros via Deus Ex … kuriose Bugs und zwei NPCs auf der Suche nach Sitzgelegenheiten Und als Bonus erklärt uns Verschwörungsexperte Christian Beuster die Wahrheit hinter zwei Verschwörungsmythen des Spiels.
Gwern is a pseudonymous researcher and writer. He was one of the first people to see LLM scaling coming. If you've read his blog, you know he's one of the most interesting polymathic thinkers alive.In order to protect Gwern's anonymity, I proposed interviewing him in person, and having my friend Chris Painter voice over his words after. This amused him enough that he agreed.After the episode, I convinced Gwern to create a donation page where people can help sustain what he's up to. Please go here to contribute.Read the full transcript here.Sponsors:* Jane Street is looking to hire their next generation of leaders. Their deep learning team is looking for ML researchers, FPGA programmers, and CUDA programmers. Summer internships are open - if you want to stand out, take a crack at their new Kaggle competition. To learn more, go here: https://jane-st.co/dwarkesh* Turing provides complete post-training services for leading AI labs like OpenAI, Anthropic, Meta, and Gemini. They specialize in model evaluation, SFT, RLHF, and DPO to enhance models' reasoning, coding, and multimodal capabilities. Learn more at turing.com/dwarkesh.* This episode is brought to you by Stripe, financial infrastructure for the internet. Millions of companies from Anthropic to Amazon use Stripe to accept payments, automate financial processes and grow their revenue.If you're interested in advertising on the podcast, check out this page.Timestamps00:00:00 - Anonymity00:01:09 - Automating Steve Jobs00:04:38 - Isaac Newton's theory of progress00:06:36 - Grand theory of intelligence00:10:39 - Seeing scaling early00:21:04 - AGI Timelines00:22:54 - What to do in remaining 3 years until AGI00:26:29 - Influencing the shoggoth with writing00:30:50 - Human vs artificial intelligence00:33:52 - Rabbit holes00:38:48 - Hearing impairment00:43:00 - Wikipedia editing00:47:43 - Gwern.net00:50:20 - Counterfactual careers00:54:30 - Borges & literature01:01:32 - Gwern's intelligence and process01:11:03 - A day in the life of Gwern01:19:16 - Gwern's finances01:25:05 - The diversity of AI minds01:27:24 - GLP drugs and obesity01:31:08 - Drug experimentation01:33:40 - Parasocial relationships01:35:23 - Open rabbit holes Get full access to Dwarkesh Podcast at www.dwarkeshpatel.com/subscribe
Nobody wants to miss out on the next big thing. But “the next big thing” may, in fact, be nothing more than a dud. How can investors find the happy medium FOMO and foresight? Senior Fool Analyst Asit Sharma joins Ricky Mulvey for a conversation on the different reasons why investors buy stocks. They also discuss: What we can learn from King Charles' portfolio. The math of winners vs. losers. How to think about expected value. Tickers mentioned: SFT, PLTR, BTC, CRSP, RKLB, USHY Host: Ricky Mulvey Guest: Asit Sharma Producer: Mary Long Engineer: Tim Sparks, Austin Morgan Learn more about your ad choices. Visit megaphone.fm/adchoices