Podcasts about ai systems

  • 319PODCASTS
  • 443EPISODES
  • 40mAVG DURATION
  • 5WEEKLY NEW EPISODES
  • Jun 21, 2025LATEST

POPULARITY

20172018201920202021202220232024


Best podcasts about ai systems

Latest podcast episodes about ai systems

Cyber Security Today
AI Vulnerabilities and the Gentle Singularity: A Deep Dive with Project Synapse

Cyber Security Today

Play Episode Listen Later Jun 21, 2025 60:59 Transcription Available


In this thought-provoking episode of Project Synapse, host Jim and his friends Marcel Gagne and John Pinard delve into the complexities of artificial intelligence, especially in the context of cybersecurity. The discussion kicks off by revisiting a blog post by Sam Altman about reaching a 'Gentle Singularity' in AI development, where the progress towards artificial superintelligence seems inevitable. They explore the idea of AI surpassing human intelligence and the implications of machines learning to write their own code. Throughout their engaging conversation, they emphasize the need to integrate security into AI systems from the start, rather than as an afterthought, citing recent vulnerabilities like Echo Leak and Microsoft Copilot's Zero Click vulnerability. Derailing into stories from the past and pondering philosophical questions, they wrap up by urging for a balanced approach where speed and thoughtful planning coexist, and to prioritize human welfare in technological advancements. This episode serves as a captivating blend of storytelling, technical insights, and ethical debates. 00:00 Introduction to Project Synapse 00:38 AI Vulnerabilities and Cybersecurity Concerns 02:22 The Gentle Singularity and AI Evolution 04:54 Human and AI Intelligence: A Comparison 07:05 AI Hallucinations and Emotional Intelligence 12:10 The Future of AI and Its Limitations 27:53 Security Flaws in AI Systems 30:20 The Need for Robust AI Security 32:22 The Ubiquity of AI in Modern Society 32:49 Understanding Neural Networks and Model Security 34:11 Challenges in AI Security and Human Behavior 36:45 The Evolution of Steganography and Prompt Injection 39:28 AI in Automation and Manufacturing 40:49 Crime as a Business and Security Implications 42:49 Balancing Speed and Security in AI Development 53:08 Corporate Responsibility and Ethical Considerations 57:31 The Future of AI and Human Values

Artificial Intelligence in Industry with Daniel Faggella
Building AI Systems That Think Like Scientists in Life Sciences - Annabel Romero of Deloitte

Artificial Intelligence in Industry with Daniel Faggella

Play Episode Listen Later Jun 18, 2025 31:01


Today's guest is Annabel Romero, Specialist Leader focusing on AI for Drug Discovery at Deloitte and a structural biologist by training. Deloitte is a global consulting firm known for its work in digital transformation, data strategy, and AI adoption across regulated industries. Annabel joins Emerj Editorial Director Matthew DeMello to explore how AI systems are being designed to think more like scientists—particularly in protein modeling and life sciences research. She shares how tools like AlphaFold and large language models are accelerating drug targeting, predicting allergen cross-reactivity, and translating learnings from human biology to agricultural innovation. This episode is sponsored by Deloitte. Want to share your AI adoption story with executive peers? Click emerj.com/expert2 for more information and to be a potential future guest on the ‘AI in Business' podcast!

MONEY FM 89.3 - The Breakfast Huddle with Elliott Danker, Manisha Tank and Finance Presenter Ryan Huang
Mind Your Business: The key to building resilient AI systems with strong governance

MONEY FM 89.3 - The Breakfast Huddle with Elliott Danker, Manisha Tank and Finance Presenter Ryan Huang

Play Episode Listen Later Jun 12, 2025 13:17


As AI rapidly shifts from experimental to essential, organizations around the world are racing to embed it into the very fabric of their operations. But beneath the excitement lies a growing concern: how do we scale AI responsibly—without compromising data quality, security, or public trust? The Breakfast Show explores what it really takes to future-proof businesses in an AI-driven world with David Irecki, Chief Technology Officer for Asia Pacific & Japan at Boomi, as they unpack how organizations can prepare for the future of AI-driven integration and automation—while staying grounded in trust, transparency, and control.See omnystudio.com/listener for privacy information.

Claims Game Podcast with Vince Perri
How Jeremy Kloter Scaled to 200+ Rental Doors Using AI, Systems & Leadership w/ Jeremy Kloter

Claims Game Podcast with Vince Perri

Play Episode Listen Later Jun 10, 2025 56:32


Beyond The Claim Podcast #014 – Featuring Jeremy Kloter Unlock the Systems Behind 200+ Rental Doors & Real Estate Growth In this episode of Beyond The Claim, Vince Perri sits down with Jeremy Kloter, Marine Corps veteran and Founder of Out Fast Property Management, to break down how he scaled to over 200 rental units using smart systems, AI tools, and leadership strategies. Jeremy shares the secrets behind: - Building a real estate business with operational excellence - Integrating AI and automation to manage properties efficiently - How his military background shaped his leadership style - Strategic use of tech like Notion, monday.com, and SOPs - Why property management is the untapped cash flow machine in 2025 - What every investor needs to know about taxes, insurance, and ROI today

The Heidrick & Struggles Leadership Podcast
Building trust in AI systems and AI systems that deserve our trust: A conversation with Miriam Vogel, president and CEO of EqualAI

The Heidrick & Struggles Leadership Podcast

Play Episode Listen Later Jun 10, 2025 26:49


In this interview, as part of our ongoing series examining how leaders across functions are managing and incorporating AI, Heidrick & Struggles' Julian Ha spoke to Miriam Vogel the president and CEO of EqualAI, which is a nonprofit organization created to promote artificial intelligence governance.Miriam, who has extensive experience working with C-suites and boards and the government to help them establish best practices for legal regulatory compliance, as well as operationalizing those best practices, discusses the ways that leaders of all kinds have become much more AI literate with respect to potential harms and litigation risks, and how they are mitigating liability as well as avoiding those harm to companies, individuals, and communities. She also shares what she's excited about when it comes to the future of AI, what it means to her to lead in an AI-enabled world, and how she believes that leaders must balance guardrails and innovation, so that their organizations remain competitive, ensure there's trust in AI systems, and that those AI systems deserve our trust. Hosted on Acast. See acast.com/privacy for more information.

MLOps.community
Product Metrics are LLM Evals // Raza Habib CEO of Humanloop // #320

MLOps.community

Play Episode Listen Later Jun 3, 2025 53:06


Raza Habib, the CEO of LLM Eval platform Humanloop, talks to us about how to make your AI products more accurate and reliable by shortening the feedback loop of your evals. Quickly iterating on prompts and testing what works, along with some of his favorite Dario from Anthropic AI Quotes.// BioRaza is the CEO and Co-founder at Humanloop. He has a PhD in Machine Learning from UCL, was the founding engineer of Monolith AI, and has built speech systems at Google. For the last 4 years, he has led Humanloop and supported leading technology companies such as Duolingo, Vanta, and Gusto to build products with large language models. Raza was featured in the Forbes 30 Under 30 technology list in 2022, and Sifted recently named him one of the most influential Gen AI founders in Europe.// Related LinksWebsites: https://humanloop.com~~~~~~~~ ✌️Connect With Us ✌️ ~~~~~~~Catch all episodes, blogs, newsletters, and more: https://go.mlops.community/TYExploreMLOps Swag/Merch: [https://shop.mlops.community/]Connect with Demetrios on LinkedIn: /dpbrinkmConnect with Raza on LinkedIn: /humanloop-razaTimestamps:[00:00] Cracking Open System Failures and How We Fix Them[05:44] LLMs in the Wild — First Steps and Growing Pains[08:28] Building the Backbone of Tracing and Observability[13:02] Tuning the Dials for Peak Model Performance[13:51] From Growing Pains to Glowing Gains in AI Systems[17:26] Where Prompts Meet Psychology and Code[22:40] Why Data Experts Deserve a Seat at the Table[24:59] Humanloop and the Art of Configuration Taming[28:23] What Actually Matters in Customer-Facing AI[33:43] Starting Fresh with Private Models That Deliver[34:58] How LLM Agents Are Changing the Way We Talk[39:23] The Secret Lives of Prompts Inside Frameworks[42:58] Streaming Showdowns — Creativity vs. Convenience[46:26] Meet Our Auto-Tuning AI Prototype[49:25] Building the Blueprint for Smarter AI[51:24] Feedback Isn't Optional — It's Everything

AI Unraveled: Latest AI News & Trends, Master GPT, Gemini, Generative AI, LLMs, Prompting, GPT Store

Master the future of autonomous systems with this essential guide to the Top 20 Agentic Patterns. Designed for AI engineers and system architects, this hands-on book provides practical blueprints for building intelligent, adaptive, and autonomous agents. Learn how to apply proven design patterns—from perception-action loops to belief-desire-intention (BDI) models—to develop powerful agentic systems that reason, learn, and act.

How to Scale an Agency
How to Future-Proof Your Digital Marketing Agency with AI: Systems, Tools & Strategies That Scale (Jordan Ross Solo)

How to Scale an Agency

Play Episode Listen Later May 26, 2025 9:46


In this episode, Jordan Ross delivers a clear message: AI is here, but most businesses aren't ready. He explains why agencies and entrepreneurs need to stop treating their businesses like marketing funnels and start thinking like architects of long-lasting infrastructure. Rushed systems and short-term thinking are costing founders time, money, and relevance. Jordan shares how to future-proof your operations by designing them for scale, sustainability, and AI integration—before it's too late.⏱ Chapters:– The Urgency of AI Preparedness– Structuring Your Business for AI Success– Rethinking Operational Infrastructure– The Importance of Long-Term PlanningTo learn more go to 8figureagency.co

AI + a16z
Building AI Systems You Can Trust

AI + a16z

Play Episode Listen Later May 23, 2025 47:40


In this episode of AI + a16z, Distributional cofounder and CEO Scott Clark, and a16z partner Matt Bornstein, explore why building trust in AI systems matters more than just optimizing performance metrics. From understanding the hidden complexities of generative AI behavior to addressing the challenges of reliability and consistency, they discuss how to confidently deploy AI in production. Why is trust becoming a critical factor in enterprise AI adoption? How do traditional performance metrics fail to capture crucial behavioral nuances in generative AI systems? Scott and Matt dive into these questions, examining non-deterministic outcomes, shifting model behaviors, and the growing importance of robust testing frameworks. Among other topics, they cover: The limitations of conventional AI evaluation methods and the need for behavioral testing. How centralized AI platforms help enterprises manage complexity and ensure responsible AI use. The rise of "shadow AI" and its implications for security and compliance. Practical strategies for scaling AI confidently from prototypes to real-world applications.Follow everyone:Scott ClarkDistributionalMatt BornsteinDerrick Harris Check out everything a16z is doing with artificial intelligence here, including articles, projects, and more podcasts.

Twisted Younginz
Final Destination | Real ID, Ai System leads to Mark of The Beast, Converting to Christianity #watch

Twisted Younginz

Play Episode Listen Later May 23, 2025 107:03


Have you examined the book of revelations and disected the events, signs, and warnings that Jesus said to look out for?? If you haven't we got you covered, check out end time news, world news, and if you are not saved please reconsider and give your life to Jesus.If you would like to support our movement please send all donations to our cash app $Twistedyounginz (your support is greatly appreciated)IG@TwistedYounginzTikTok@TwistedYounginzrumble @Twistedyounginz we are available on streaming platform

This Day in AI Podcast
The Future of AI Systems: EP99.04-PREVIEW

This Day in AI Podcast

Play Episode Listen Later May 16, 2025 85:37


Join Simtheory: https://simtheory.aiGet an AI workspace for your team: https://simtheory.ai/workspace/team/---CHAPTERS:00:00 - Will Chris Lose His Bet?04:48 - Google's 2.5 Gemini Preview Update12:44 - Future AI Systems Discussion: Skills, MCPs & A2A47:02 - Will AI Systems become walled gardens?55:13 - Do Organizations That Own Data Build MCPs & Agents? Is This The New SaaS?1:17:45 - Can we improve RAG with tool calling and stop hallucinations?---Thanks for listening. If you like chatting about AI consider joining our active Discord community: https://thisdayinai.com.

Daily Tech Headlines
DeepMind Unveils AI System AlphaEvolve – DTH

Daily Tech Headlines

Play Episode Listen Later May 14, 2025


Warners Bros Discovery goes back to HBO Max, Sony warns of PS5 price hikes, Google replaces “I'm feeling lucky” with AI mode. MP3 Please SUBSCRIBE HERE for free or get DTNS Live ad-free. A special thanks to all our supporters–without you, none of this would be possible. If you enjoy what you see you canContinue reading "DeepMind Unveils AI System AlphaEvolve – DTH"

Finding Genius Podcast
Autonomous AI Systems: Entering A New Era Of Technology With Shirish Nimgaonkar

Finding Genius Podcast

Play Episode Listen Later May 13, 2025 21:01


What is “self-healing AI?” How do prediction and personalization deliver a superior ROI and enhanced user experience? In this episode, we are joined by Shirish Nimgaonkar to dive into this intriguing and revolutionary topic… Shirish is an entrepreneur, advisor, and investor who focuses his skills on software and AI. He is currently the Founder and CEO of eBliss, a revolutionary AI-driven autonomous end-user computing platform dedicated to streamlining the digital workplace – boosting operational performance, anticipating and resolving IT issues, and elevating both productivity and user satisfaction. Hit play to find out: How businesses can reduce operational costs using personalized AI. The problems that exist within different categories of devices. The ways that predictive analytics can improve productivity.  Industries that benefit from AI solutions. Shirish is a seasoned tech leader who has led and scaled high-growth software companies. He has held leadership roles at several PE and VC-backed tech firms and previously founded and led the South Asia group at a global investment bank, where he oversaw over 30 client acquisitions. Currently, he serves as an Entrepreneur in Residence at Harvard Business School and advises multiple startups. Shirish holds degrees from IIT Bombay, Stanford, and Harvard Business School. You can find out more about Shirish and his work here! Episode also available on Apple Podcasts: https://apple.co/30PvU9C

Voices of Search // A Search Engine Optimization (SEO) & Content Marketing Podcast
Will AI Systems Be Able To Fully Diagnose Traffic Drops And Identify Fundamental Issues?

Voices of Search // A Search Engine Optimization (SEO) & Content Marketing Podcast

Play Episode Listen Later May 10, 2025 3:09


Can AI fully diagnose SEO traffic drops? Sam Torres from Gray Dot Co examines the limitations of AI in technical SEO diagnostics. She explains why AI can identify traffic declines but struggles to pinpoint fundamental issues, emphasizing that most SEO problems stem from multiple factors rather than single causes. Torres highlights the importance of human experience in comprehensive troubleshooting and avoiding the "one solution" trap that both AI and junior SEOs often fall into. Show NotesConnect With: Sam Torres: Website // LinkedInThe Voices of Search Podcast: Email // LinkedIn // TwitterBenjamin Shapiro: Website // LinkedIn // TwitterSee Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

The AI Fundamentalists
Principles, agents, and the chain of accountability in AI systems

The AI Fundamentalists

Play Episode Listen Later May 8, 2025 46:26 Transcription Available


Dr. Michael Zargham provides a systems engineering perspective on AI agents, emphasizing accountability structures and the relationship between principals who deploy agents and the agents themselves. In this episode, he brings clarity to the often misunderstood concept of agents in AI by grounding them in established engineering principles rather than treating them as mysterious or elusive entities.Show highlights• Agents should be understood through the lens of the principal-agent relationship, with clear lines of accountability• True validation of AI systems means ensuring outcomes match intentions, not just optimizing loss functions• LLMs by themselves are "high-dimensional word calculators," not agents - agents are more complex systems with LLMs as components• Guardrails provide deterministic constraints ("musts" or "shalls") versus constitutional AI's softer guidance ("shoulds")• Systems engineering approaches from civil engineering and materials science offer valuable frameworks for AI development• Authority and accountability must align - people shouldn't be held responsible for systems they don't have authority to control• The transition from static input-output to closed-loop dynamical systems represents the shift toward truly agentic behavior• Robust agent systems require both exploration (lab work) and exploitation (hardened deployment) phases with different standardsExplore Dr. Zargham's workProtocols and Institutions (Feb 27, 2025)Comments Submitted by BlockScience, University of Washington APL Information Risk and Synthetic Intelligence Research Initiative (IRSIRI), Cognitive Security and Education Forum (COGSEC), and the Active Inference Institute (AII) to the Networking and Information Technology Research and Development National Coordination Office's Request for Comment on The Creation of a National Digital Twins R&D Strategic Plan NITRD-2024-13379 (Aug 8, 2024)What did you think? Let us know.Do you have a question or a discussion topic for the AI Fundamentalists? Connect with them to comment on your favorite topics: LinkedIn - Episode summaries, shares of cited articles, and more. YouTube - Was it something that we said? Good. Share your favorite quotes. Visit our page - see past episodes and submit your feedback! It continues to inspire future episodes.

Strap on your Boots!
Episode 301: Zero to CEO: Revolutionize Your IT with Autonomous AI Systems with Shirish Nimgaonkar

Strap on your Boots!

Play Episode Listen Later May 5, 2025 15:36


In this episode of Zero to CEO, I speak with Harvard serial founder Shirish Nimgaonkar about how his company, eBlissAI, is leading a new era of agentic AI through self-healing, autonomous systems. Shirish explains why today's enterprise IT systems are falling short and how eBlissAI's platform is transforming endpoint management by combining real-time analytics, predictive insights, and intelligent automation. We discuss how businesses can drastically reduce operational costs, improve user productivity, and gain a superior ROI by adopting AI that not only reacts but predicts and personalizes. With enterprise tech at a tipping point, this episode explores why now is the critical moment to embrace truly autonomous solutions.

AI at Scale
Renée Cummings: Designing AI systems for high-stakes decisions

AI at Scale

Play Episode Listen Later May 5, 2025 27:16


How to make the most competitive AI systems? Renée Cummings, a renowned Professor of Practice in Data Science at the University of Virginia and a leading voice in AI ethics, talks about the importance of trust, transparency, and accountability in AI development, making a strong case for why these elements are essential for businesses aiming to leverage AI effectively. Furthermore, Renée shares her expertise on the intersection of AI and data justice, where AI is applied in high-stakes decision-making process.

Middle Ground with JLE
“Journey Of HolyGround Master Blueprint” with Special Guest Founder/CEO Brian Bontomase

Middle Ground with JLE

Play Episode Listen Later May 4, 2025 38:14


Middle Ground with JLE L.L.C. “Where We Treat You Like Family” Automating Kingdom Growth, AI Systems for Faith-Driven leaders, Author, and HolyGround Master Blueprint Founder/CEO Brian (Kingdom Wealth Generator) Bontomase as he shares his journey from a successful Tree Company to losing it all and finding his true purpose.

The Data Exchange with Ben Lorica
Beyond the Demo: Building AI Systems That Actually Work

The Data Exchange with Ben Lorica

Play Episode Listen Later May 1, 2025 27:36


Hamel Husain is the founder of Parlance Labs. He discusses how successful AI implementation requires fundamental data science skills often overlooked in current educational resources that focus too heavily on tools and frameworks. Subscribe to the Gradient Flow Newsletter

UM HELLO?
You Might Also Like: On Purpose with Jay Shetty

UM HELLO?

Play Episode Listen Later Apr 24, 2025


Introducing Deepak Chopra: How to Ask AI The RIGHT Questions To Grow, Heal, and Live a More Fulfilled Life from On Purpose with Jay Shetty.Follow the show: On Purpose with Jay ShettyDo you ever feel like you're just going through the motions? Do you ever feel lonely even when you're around people? Today, Jay welcomes back the legendary Deepak Chopra after six years, to discuss the unexpected intersection of spirituality and artificial intelligence. Together they unpack the beautiful blend of ancient wisdom and cutting-edge innovation. Jay reflects on his first meeting with Deepak and how their bond has evolved over the years. The conversation quickly flows into one of the most fascinating questions of our time: Can AI actually support our spiritual growth, rather than distract from it? Deepak shares his insights on the mysteries of the universe, the unseen forces that shape our reality, and why he believes consciousness, not matter, is the foundation of everything. Jay and Deepak break down the fears surrounding AI, from misinformation to war, and why it’s so important for us to grow spiritually alongside our technology. They explore how storytelling, creativity, and love are deeply human qualities that no machine can replicate, and how we can use these gifts to shape a better, more compassionate world. In this interview, you'll learn: How to Ask AI Better Questions to Unlock Wisdom How to Use Technology to Support Emotional and Physical Healing How to Find Purpose Through the Art of Storytelling How to Train AI to Reflect Inclusivity and Diversity How to Embrace Consciousness as the Foundation of Reality How to Create a Sacred Relationship with Technology Deepak reminds us that the answers we’re looking for often begin with a better question. And with the right mindset, even the most advanced technology can become a tool for inner transformation and collective healing. With Love and Gratitude, Jay Shetty Join over 750,000 people to receive my most transformative wisdom directly in your inbox every single week with my free newsletter. Subscribe here. Join Jay for his first ever, On Purpose Live Tour! Tickets are on sale now. Hope to see you there! What We Discuss: 00:00 Intro 01:41 What If the Universe Is Just a Giant Digital Simulation? 12:36 How to Train AI to Unlock Ancient and Hidden Knowledge 13:39 Blending AI and Spirituality to Understand Consciousness 18:23 Could AI Really Lead to Human Extinction? 22:34 What’s Actually Holding Humanity Back From Progress? 23:37 How the Human Brain Transformed Over Time 26:11 The 2 Things That Set Humans Apart From All Other Species 27:16 Can Technology Lead Us to True Peace and Prosperity? 28:10 Will AI Replace Our Jobs or Unlock Human Creativity? 30:46 Do You Think AI Can Ever Have a Soul? 31:45 The Gender and Racial Bias Hidden in AI Systems 32:33 How to Build More Inclusive and Equitable AI Models 33:32 Why a Shared Vision Can Solve Any Problem We Face 36:13 Would You Trust AI to Know You Personally? 36:57 How You can Use AI to Get Better Sleep 37:29 Can AI Actually Give You Good Relationship Advice? 38:07 How AI Can Help You Find and Nurture Love 38:29 Why Personal Growth Solutions Should Never Be Generic 39:41 Your DNA Holds the Footprints of Human History 42:33 Rethinking the Big Bang: What Science Still Can’t Explain 44:31 Is Everything You See Just a Projection? 47:48 Why Fear of the Unknown Limits Our Growth 48:52 Want Better Answers? Ask Better Questions 51:41 The True Secret to Longevity Isn’t What You Think 54:30 How Your Brain Turns Experience Into Reality 55:08 Why Consciousness Is Still Life’s Greatest Mystery 56:28 The First Question You Should Always Ask AI 58:38 How ChatGPT Can Spark Deeper, More Intelligent Questions Episode Resources: Deepak Chopra | Website Deepak Chopra | Instagram Deepak Chopra | TikTok Deepak Chopra | Facebook Deepak Chopra | YouTube Deepak Chopra | X Digital Dharma: How AI Can Elevate Spiritual Intelligence and Personal Well-BeingSee omnystudio.com/listener for privacy information.DISCLAIMER: Please note, this is an independent podcast episode not affiliated with, endorsed by, or produced in conjunction with the host podcast feed or any of its media entities. The views and opinions expressed in this episode are solely those of the creators and guests. For any concerns, please reach out to team@podroll.fm.

This Day in AI Podcast
From AI Models to AI Systems and The Future of Vibe Gaming: An Average Talk

This Day in AI Podcast

Play Episode Listen Later Apr 23, 2025 52:50


Don't Let It Stu
You Might Also Like: On Purpose with Jay Shetty

Don't Let It Stu

Play Episode Listen Later Apr 22, 2025


Introducing Deepak Chopra: How to Ask AI The RIGHT Questions To Grow, Heal, and Live a More Fulfilled Life from On Purpose with Jay Shetty.Follow the show: On Purpose with Jay ShettyDo you ever feel like you're just going through the motions? Do you ever feel lonely even when you're around people? Today, Jay welcomes back the legendary Deepak Chopra after six years, to discuss the unexpected intersection of spirituality and artificial intelligence. Together they unpack the beautiful blend of ancient wisdom and cutting-edge innovation. Jay reflects on his first meeting with Deepak and how their bond has evolved over the years. The conversation quickly flows into one of the most fascinating questions of our time: Can AI actually support our spiritual growth, rather than distract from it? Deepak shares his insights on the mysteries of the universe, the unseen forces that shape our reality, and why he believes consciousness, not matter, is the foundation of everything. Jay and Deepak break down the fears surrounding AI, from misinformation to war, and why it’s so important for us to grow spiritually alongside our technology. They explore how storytelling, creativity, and love are deeply human qualities that no machine can replicate, and how we can use these gifts to shape a better, more compassionate world. In this interview, you'll learn: How to Ask AI Better Questions to Unlock Wisdom How to Use Technology to Support Emotional and Physical Healing How to Find Purpose Through the Art of Storytelling How to Train AI to Reflect Inclusivity and Diversity How to Embrace Consciousness as the Foundation of Reality How to Create a Sacred Relationship with Technology Deepak reminds us that the answers we’re looking for often begin with a better question. And with the right mindset, even the most advanced technology can become a tool for inner transformation and collective healing. With Love and Gratitude, Jay Shetty Join over 750,000 people to receive my most transformative wisdom directly in your inbox every single week with my free newsletter. Subscribe here. Join Jay for his first ever, On Purpose Live Tour! Tickets are on sale now. Hope to see you there! What We Discuss: 00:00 Intro 01:41 What If the Universe Is Just a Giant Digital Simulation? 12:36 How to Train AI to Unlock Ancient and Hidden Knowledge 13:39 Blending AI and Spirituality to Understand Consciousness 18:23 Could AI Really Lead to Human Extinction? 22:34 What’s Actually Holding Humanity Back From Progress? 23:37 How the Human Brain Transformed Over Time 26:11 The 2 Things That Set Humans Apart From All Other Species 27:16 Can Technology Lead Us to True Peace and Prosperity? 28:10 Will AI Replace Our Jobs or Unlock Human Creativity? 30:46 Do You Think AI Can Ever Have a Soul? 31:45 The Gender and Racial Bias Hidden in AI Systems 32:33 How to Build More Inclusive and Equitable AI Models 33:32 Why a Shared Vision Can Solve Any Problem We Face 36:13 Would You Trust AI to Know You Personally? 36:57 How You can Use AI to Get Better Sleep 37:29 Can AI Actually Give You Good Relationship Advice? 38:07 How AI Can Help You Find and Nurture Love 38:29 Why Personal Growth Solutions Should Never Be Generic 39:41 Your DNA Holds the Footprints of Human History 42:33 Rethinking the Big Bang: What Science Still Can’t Explain 44:31 Is Everything You See Just a Projection? 47:48 Why Fear of the Unknown Limits Our Growth 48:52 Want Better Answers? Ask Better Questions 51:41 The True Secret to Longevity Isn’t What You Think 54:30 How Your Brain Turns Experience Into Reality 55:08 Why Consciousness Is Still Life’s Greatest Mystery 56:28 The First Question You Should Always Ask AI 58:38 How ChatGPT Can Spark Deeper, More Intelligent Questions Episode Resources: Deepak Chopra | Website Deepak Chopra | Instagram Deepak Chopra | TikTok Deepak Chopra | Facebook Deepak Chopra | YouTube Deepak Chopra | X Digital Dharma: How AI Can Elevate Spiritual Intelligence and Personal Well-BeingSee omnystudio.com/listener for privacy information.DISCLAIMER: Please note, this is an independent podcast episode not affiliated with, endorsed by, or produced in conjunction with the host podcast feed or any of its media entities. The views and opinions expressed in this episode are solely those of the creators and guests. For any concerns, please reach out to team@podroll.fm.

My Friend, My Soulmate, My Podcast
You Might Also Like: On Purpose with Jay Shetty

My Friend, My Soulmate, My Podcast

Play Episode Listen Later Apr 22, 2025


Introducing Deepak Chopra: How to Ask AI The RIGHT Questions To Grow, Heal, and Live a More Fulfilled Life from On Purpose with Jay Shetty.Follow the show: On Purpose with Jay ShettyDo you ever feel like you're just going through the motions? Do you ever feel lonely even when you're around people? Today, Jay welcomes back the legendary Deepak Chopra after six years, to discuss the unexpected intersection of spirituality and artificial intelligence. Together they unpack the beautiful blend of ancient wisdom and cutting-edge innovation. Jay reflects on his first meeting with Deepak and how their bond has evolved over the years. The conversation quickly flows into one of the most fascinating questions of our time: Can AI actually support our spiritual growth, rather than distract from it? Deepak shares his insights on the mysteries of the universe, the unseen forces that shape our reality, and why he believes consciousness, not matter, is the foundation of everything. Jay and Deepak break down the fears surrounding AI, from misinformation to war, and why it’s so important for us to grow spiritually alongside our technology. They explore how storytelling, creativity, and love are deeply human qualities that no machine can replicate, and how we can use these gifts to shape a better, more compassionate world. In this interview, you'll learn: How to Ask AI Better Questions to Unlock Wisdom How to Use Technology to Support Emotional and Physical Healing How to Find Purpose Through the Art of Storytelling How to Train AI to Reflect Inclusivity and Diversity How to Embrace Consciousness as the Foundation of Reality How to Create a Sacred Relationship with Technology Deepak reminds us that the answers we’re looking for often begin with a better question. And with the right mindset, even the most advanced technology can become a tool for inner transformation and collective healing. With Love and Gratitude, Jay Shetty Join over 750,000 people to receive my most transformative wisdom directly in your inbox every single week with my free newsletter. Subscribe here. Join Jay for his first ever, On Purpose Live Tour! Tickets are on sale now. Hope to see you there! What We Discuss: 00:00 Intro 01:41 What If the Universe Is Just a Giant Digital Simulation? 12:36 How to Train AI to Unlock Ancient and Hidden Knowledge 13:39 Blending AI and Spirituality to Understand Consciousness 18:23 Could AI Really Lead to Human Extinction? 22:34 What’s Actually Holding Humanity Back From Progress? 23:37 How the Human Brain Transformed Over Time 26:11 The 2 Things That Set Humans Apart From All Other Species 27:16 Can Technology Lead Us to True Peace and Prosperity? 28:10 Will AI Replace Our Jobs or Unlock Human Creativity? 30:46 Do You Think AI Can Ever Have a Soul? 31:45 The Gender and Racial Bias Hidden in AI Systems 32:33 How to Build More Inclusive and Equitable AI Models 33:32 Why a Shared Vision Can Solve Any Problem We Face 36:13 Would You Trust AI to Know You Personally? 36:57 How You can Use AI to Get Better Sleep 37:29 Can AI Actually Give You Good Relationship Advice? 38:07 How AI Can Help You Find and Nurture Love 38:29 Why Personal Growth Solutions Should Never Be Generic 39:41 Your DNA Holds the Footprints of Human History 42:33 Rethinking the Big Bang: What Science Still Can’t Explain 44:31 Is Everything You See Just a Projection? 47:48 Why Fear of the Unknown Limits Our Growth 48:52 Want Better Answers? Ask Better Questions 51:41 The True Secret to Longevity Isn’t What You Think 54:30 How Your Brain Turns Experience Into Reality 55:08 Why Consciousness Is Still Life’s Greatest Mystery 56:28 The First Question You Should Always Ask AI 58:38 How ChatGPT Can Spark Deeper, More Intelligent Questions Episode Resources: Deepak Chopra | Website Deepak Chopra | Instagram Deepak Chopra | TikTok Deepak Chopra | Facebook Deepak Chopra | YouTube Deepak Chopra | X Digital Dharma: How AI Can Elevate Spiritual Intelligence and Personal Well-BeingSee omnystudio.com/listener for privacy information.DISCLAIMER: Please note, this is an independent podcast episode not affiliated with, endorsed by, or produced in conjunction with the host podcast feed or any of its media entities. The views and opinions expressed in this episode are solely those of the creators and guests. For any concerns, please reach out to team@podroll.fm.

Workday Podcast
How To Build Culture, Trust, and Ethics into AI Systems - Future of Work Podcast

Workday Podcast

Play Episode Listen Later Apr 22, 2025 47:11


In this episode of the Future of Work podcast, Reggie Townsend, vice president of the data ethics practice at SAS and member of the U.S. National AI Advisory Committee, joins Kathy Pham, Workday's Vice President of Artificial Intelligence, to break down what AI readiness really means, from building cultural trust and ethical frameworks to preparing infrastructure and governance for agentic AI.

RTÉ - Morning Ireland
Irish Authors to submit petition to Government against Meta's AI system Llama 3

RTÉ - Morning Ireland

Play Episode Listen Later Apr 17, 2025 7:21


Audrey Magee, Booker prize longlisted author and Conor Kostick, author and advocacy officer at the Irish Writers Union explain why they will be protesting outside the Department of Trade over the use of their copyrighted work to train AI models.

Closers Are Losers with Jeremy Miner
Creating the AI SYSTEM That Can 10X Your High-Ticket Leads with Douglas James | Ep. 368

Closers Are Losers with Jeremy Miner

Play Episode Listen Later Apr 16, 2025 51:33


The EXACT Objection Handling Framework I Used To Earn $2.4MM/yr In Straight Commissions As A W-2 Sales Rep Link ➡️: https://nepqtraining.com/smv-yt-splt-opt-org Text me if you have any sales, persuasion, or influence questions: +1-480-637-2944 In this episode, Jeremy Miner dives deep with Douglas James, a traffic and sales expert, to discuss how AI is revolutionizing lead generation. Discover how cutting-edge technology helps businesses make smarter decisions faster and why your sales strategy might be leaving money on the table. If you're tired of wasting time with unqualified leads and want to scale your business, this episode is a must-watch! ✅ Resources: JOIN the Sales Revolution: https://www.facebook.com/groups/salesrevolutiongroup Book a "Clarity CALL": https://7thlevelhq.com/book-demo/ ✅ Connect with Me: Follow Jeremy Miner on Facebook: https://www.facebook.com/jeremy.miner.52 Follow Jeremy Miner on Instagram: https://www.instagram.com/jeremyleeminer/ Follow Jeremy Miner on LinkedIn: https://www.linkedin.com/in/jeremyleeminer/ ✅ SUBSCRIBE to My Podcast CLOSERS ARE LOSERS with Jeremy Miner: Subscribe on iTunes: https://podcasts.apple.com/us/podcast/closers-are-losers-with-jeremyminer/id1534365100 Subscribe and Review on Spotify: https://open.spotify.com/show/2kNDyUR7fz9SqBr9iGwfwV?si=uMhsOBP4S_SBaHqAFp4EGg Subscribe and Review on Stitcher: https://www.stitcher.com/show/closers-are-losers-with-jeremy-miner Subscribe and Review on Google Podcasts: https://podcasts.google.com/u/1/feed/aHR0cHM6Ly9jbG9zZXJzYXJlbG9zZXJzLmxpYnN5bi5jb20vcnNz  

Eye On A.I.
#247 Barr Moses: Why Reliable Data is Key to Building Good AI Systems

Eye On A.I.

Play Episode Listen Later Apr 13, 2025 55:36


This episode is sponsored by Netsuite by Oracle, the number one cloud financial system, streamlining accounting, financial management, inventory, HR, and more.   NetSuite is offering a one-of-a-kind flexible financing program. Head to  https://netsuite.com/EYEONAI to know more.   In this episode of Eye on AI, Craig Smith sits down with Barr Moses, Co-Founder & CEO of Monte Carlo, the pioneer of data and AI observability. Together, they explore the hidden force behind every great AI system: reliable, trustworthy data. With AI adoption soaring across industries, companies now face a critical question: Can we trust the data feeding our models? Barr unpacks why data quality is more important than ever, how observability helps detect and resolve data issues, and why clean data—not access to GPT or Claude—is the real competitive moat in AI today.   What You'll Learn in This Episode: Why access to AI models is no longer a competitive advantage How Monte Carlo helps teams monitor complex data estates in real-time The dangers of “data hallucinations” and how to prevent them Real-world examples of data failures and their impact on AI outputs The difference between data observability and explainability Why legacy methods of data review no longer work in an AI-first world Stay Updated: Craig Smith on X:https://x.com/craigss Eye on A.I. on X: https://x.com/EyeOn_AI     (00:00) Intro (01:08) How Monte Carlo Fixed Broken Data   (03:08) What Is Data & AI Observability?   (05:00) Structured vs Unstructured Data Monitoring   (08:48) How Monte Carlo Integrates Across Data Stacks (13:35) Why Clean Data Is the New Competitive Advantage   (16:57) How Monte Carlo Uses AI Internally   (19:20) 4 Failure Points: Data, Systems, Code, Models   (23:08) Can Observability Detect Bias in Data?   (26:15) Why Data Quality Needs a Modern Definition   (29:22) Explosion of Data Tools & Monte Carlo's 50+ Integrations   (33:18) Data Observability vs Explainability   (36:18) Human Evaluation vs Automated Monitoring   (39:23) What Monte Carlo Looks Like for Users   (46:03) How Fast Can You Deploy Monte Carlo?   (51:56) Why Manual Data Checks No Longer Work   (53:26) The Future of AI Depends on Trustworthy Data 

Orchestrate all the Things podcast: Connecting the Dots with George Anadiotis
The Quality Imperative: Why Leading Organizations Proactively Evaluate Software and AI Systems. Featuring Yiannis Kanellopoulos, code4thought CEO / Founder

Orchestrate all the Things podcast: Connecting the Dots with George Anadiotis

Play Episode Listen Later Apr 9, 2025 47:01


As enterprise systems shift from deterministic software to probabilistic AI, forward-thinking organizations leverage proactive quality assessment to maximize value, minimize risk and ensure regulatory compliance In today's rapidly evolving technological landscape, ensuring the quality of both traditional software and AI systems has become more critical than ever. Organizations are increasingly relying on complex digital systems to drive innovation and maintain competitive advantage, yet many struggle to effectively evaluate these systems before, during, and after deployment. Yiannis Kanellopoulos is on the forefront of software and AI system quality assessment. He is the founder and CEO of code4thought, a startup specializing in assessing large-scale software systems and AI applications. We connected to explore the challenges and opportunities of quality assessment and share insights both for developers and for people responsible for technology decisions within their organizations. Read the article published on Orchestrate all the Things here: https://linkeddataorchestration.com/2025/04/09/the-quality-imperative-why-leading-organizations-proactively-evaluate-software-and-ai-systems-and-how-you-can-too/

She Slays the Day
307 - Building Custom AI Systems to Automate and Accelerate Your Business and Life feat. Callan Faulkner

She Slays the Day

Play Episode Listen Later Apr 6, 2025 88:56


What if AI could think like you—and free up your time in the process? In this episode, Dr. Lauryn welcomes back Callan Faulkner, founder of REI Optimize and one of the most sought-after AI coaches for entrepreneurs, for a mind-expanding conversation on using AI to transform not just your business, but your entire life. From saving time and scaling smarter to aligning tech with your intuition, Callan breaks down exactly how to move from AI curiosity to awakened, empowered implementation.Together, they explore the difference between dabbling in AI and building custom systems that truly serve your vision. You'll learn how to create your own virtual sales assistant, brand copywriter, business coach, meal planner, and even spiritual guide using tools like ChatGPT and Claude. If you've been overwhelmed by the AI conversation—or just want to make it actually useful—this is your roadmap to building tools that think, sound, and act like you.Key Takeaways:AI isn't just for automation—it's a tool for alignment: Callan shares how to create AI systems that reflect your values, voice, and vision so you can scale your business and still feel like you.The difference between a prompt and a project is everything: Learn how to move beyond one-off prompts and instead build custom, reusable systems that save time and increase consistency.Your business deserves a custom AI dream team: From a sales assistant to a brand copywriter, Callan reveals the 3–5 AI projects every entrepreneur should build now.AI can support your personal growth too: Callan discusses how she uses AI as a spiritual coach, a journaling prompt generator, and even a relationship support tool.Guest Bio:Callan Faulkner is the founder of REI Optimize and one of the most sought-after AI coaches for entrepreneurs. She's helped over 500 businesses—from solo practitioners to 9-figure enterprises—use AI to scale smarter, reduce stress, and reclaim their time. Known for making complex technology feel human and actionable, Callan is passionate about building systems that support both business success and personal alignment. Her unique approach blends strategy with soul, helping leaders automate without losing their voice.Enroll in Callan's course, Foundations of AI, and transform how you work in just 10 days (or less!)Join The Effortless Business Bootcamp to automate your systems, multiply your output, and reclaim your freedom!Follow Callan: Instagram | LinkedIn Resources:For those interested in building a profitable personal brand in just two hours a week, check out Dr. Lauryn's new membership group Beyond Brick & Mortar!Sign up for the Weekly Slay newsletter!Follow She Slays and Dr. Lauryn: Website | Instagram | X | LinkedIn |

Machine Learning Street Talk
Eiso Kant (CTO poolside) - Superhuman Coding Is Coming!

Machine Learning Street Talk

Play Episode Listen Later Apr 2, 2025 96:28


Eiso Kant, CTO of poolside AI, discusses the company's approach to building frontier AI foundation models, particularly focused on software development. Their unique strategy is reinforcement learning from code execution feedback which is an important axis for scaling AI capabilities beyond just increasing model size or data volume. Kant predicts human-level AI in knowledge work could be achieved within 18-36 months, outlining poolside's vision to dramatically increase software development productivity and accessibility. SPONSOR MESSAGES:***Tufa AI Labs is a brand new research lab in Zurich started by Benjamin Crouzier focussed on o-series style reasoning and AGI. They are hiring a Chief Engineer and ML engineers. Events in Zurich. Goto https://tufalabs.ai/***Eiso Kant:https://x.com/eisokanthttps://poolside.ai/TRANSCRIPT:https://www.dropbox.com/scl/fi/szepl6taqziyqie9wgmk9/poolside.pdf?rlkey=iqar7dcwshyrpeoz0xa76k422&dl=0TOC:1. Foundation Models and AI Strategy [00:00:00] 1.1 Foundation Models and Timeline Predictions for AI Development [00:02:55] 1.2 Poolside AI's Corporate History and Strategic Vision [00:06:48] 1.3 Foundation Models vs Enterprise Customization Trade-offs2. Reinforcement Learning and Model Economics [00:15:42] 2.1 Reinforcement Learning and Code Execution Feedback Approaches [00:22:06] 2.2 Model Economics and Experimental Optimization3. Enterprise AI Implementation [00:25:20] 3.1 Poolside's Enterprise Deployment Strategy and Infrastructure [00:26:00] 3.2 Enterprise-First Business Model and Market Focus [00:27:05] 3.3 Foundation Models and AGI Development Approach [00:29:24] 3.4 DeepSeek Case Study and Infrastructure Requirements4. LLM Architecture and Performance [00:30:15] 4.1 Distributed Training and Hardware Architecture Optimization [00:33:01] 4.2 Model Scaling Strategies and Chinchilla Optimality Trade-offs [00:36:04] 4.3 Emergent Reasoning and Model Architecture Comparisons [00:43:26] 4.4 Balancing Creativity and Determinism in AI Models [00:50:01] 4.5 AI-Assisted Software Development Evolution5. AI Systems Engineering and Scalability [00:58:31] 5.1 Enterprise AI Productivity and Implementation Challenges [00:58:40] 5.2 Low-Code Solutions and Enterprise Hiring Trends [01:01:25] 5.3 Distributed Systems and Engineering Complexity [01:01:50] 5.4 GenAI Architecture and Scalability Patterns [01:01:55] 5.5 Scaling Limitations and Architectural Patterns in AI Code Generation6. AI Safety and Future Capabilities [01:06:23] 6.1 Semantic Understanding and Language Model Reasoning Approaches [01:12:42] 6.2 Model Interpretability and Safety Considerations in AI Systems [01:16:27] 6.3 AI vs Human Capabilities in Software Development [01:33:45] 6.4 Enterprise Deployment and Security ArchitectureCORE REFS (see shownotes for URLs/more refs):[00:15:45] Research demonstrating how training on model-generated content leads to distribution collapse in AI models, Ilia Shumailov et al. (Key finding on synthetic data risk)[00:20:05] Foundational paper introducing Word2Vec for computing word vector representations, Tomas Mikolov et al. (Seminal NLP technique)[00:22:15] OpenAI O3 model's breakthrough performance on ARC Prize Challenge, OpenAI (Significant AI reasoning benchmark achievement)[00:22:40] Seminal paper proposing a formal definition of intelligence as skill-acquisition efficiency, François Chollet (Influential AI definition/philosophy)[00:30:30] Technical documentation of DeepSeek's V3 model architecture and capabilities, DeepSeek AI (Details on a major new model)[00:34:30] Foundational paper establishing optimal scaling laws for LLM training, Jordan Hoffmann et al. (Key paper on LLM scaling)[00:45:45] Seminal essay arguing that scaling computation consistently trumps human-engineered solutions in AI, Richard S. Sutton (Influential "Bitter Lesson" perspective)

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

If you're in SF: Join us for the Claude Plays Pokemon hackathon this Sunday!If you're not: Fill out the 2025 State of AI Eng survey for $250 in Amazon cards!We are SO excited to share our conversation with Dharmesh Shah, co-founder of HubSpot and creator of Agent.ai.A particularly compelling concept we discussed is the idea of "hybrid teams" - the next evolution in workplace organization where human workers collaborate with AI agents as team members. Just as we previously saw hybrid teams emerge in terms of full-time vs. contract workers, or in-office vs. remote workers, Dharmesh predicts that the next frontier will be teams composed of both human and AI members. This raises interesting questions about team dynamics, trust, and how to effectively delegate tasks between human and AI team members.The discussion of business models in AI reveals an important distinction between Work as a Service (WaaS) and Results as a Service (RaaS), something Dharmesh has written extensively about. While RaaS has gained popularity, particularly in customer support applications where outcomes are easily measurable, Dharmesh argues that this model may be over-indexed. Not all AI applications have clearly definable outcomes or consistent economic value per transaction, making WaaS more appropriate in many cases. This insight is particularly relevant for businesses considering how to monetize AI capabilities.The technical challenges of implementing effective agent systems are also explored, particularly around memory and authentication. Shah emphasizes the importance of cross-agent memory sharing and the need for more granular control over data access. He envisions a future where users can selectively share parts of their data with different agents, similar to how OAuth works but with much finer control. This points to significant opportunities in developing infrastructure for secure and efficient agent-to-agent communication and data sharing.Other highlights from our conversation* The Evolution of AI-Powered Agents – Exploring how AI agents have evolved from simple chatbots to sophisticated multi-agent systems, and the role of MCPs in enabling that.* Hybrid Digital Teams and the Future of Work – How AI agents are becoming teammates rather than just tools, and what this means for business operations and knowledge work.* Memory in AI Agents – The importance of persistent memory in AI systems and how shared memory across agents could enhance collaboration and efficiency.* Business Models for AI Agents – Exploring the shift from software as a service (SaaS) to work as a service (WaaS) and results as a service (RaaS), and what this means for monetization.* The Role of Standards Like MCP – Why MCP has been widely adopted and how it enables agent collaboration, tool use, and discovery.* The Future of AI Code Generation and Software Engineering – How AI-assisted coding is changing the role of software engineers and what skills will matter most in the future.* Domain Investing and Efficient Markets – Dharmesh's approach to domain investing and how inefficiencies in digital asset markets create business opportunities.* The Philosophy of Saying No – Lessons from "Sorry, You Must Pass" and how prioritization leads to greater productivity and focus.Timestamps* 00:00 Introduction and Guest Welcome* 02:29 Dharmesh Shah's Journey into AI* 05:22 Defining AI Agents* 06:45 The Evolution and Future of AI Agents* 13:53 Graph Theory and Knowledge Representation* 20:02 Engineering Practices and Overengineering* 25:57 The Role of Junior Engineers in the AI Era* 28:20 Multi-Agent Systems and MCP Standards* 35:55 LinkedIn's Legal Battles and Data Scraping* 37:32 The Future of AI and Hybrid Teams* 39:19 Building Agent AI: A Professional Network for Agents* 40:43 Challenges and Innovations in Agent AI* 45:02 The Evolution of UI in AI Systems* 01:00:25 Business Models: Work as a Service vs. Results as a Service* 01:09:17 The Future Value of Engineers* 01:09:51 Exploring the Role of Agents* 01:10:28 The Importance of Memory in AI* 01:11:02 Challenges and Opportunities in AI Memory* 01:12:41 Selective Memory and Privacy Concerns* 01:13:27 The Evolution of AI Tools and Platforms* 01:18:23 Domain Names and AI Projects* 01:32:08 Balancing Work and Personal Life* 01:35:52 Final Thoughts and ReflectionsTranscriptAlessio [00:00:04]: Hey everyone, welcome back to the Latent Space podcast. This is Alessio, partner and CTO at Decibel Partners, and I'm joined by my co-host Swyx, founder of Small AI.swyx [00:00:12]: Hello, and today we're super excited to have Dharmesh Shah to join us. I guess your relevant title here is founder of Agent AI.Dharmesh [00:00:20]: Yeah, that's true for this. Yeah, creator of Agent.ai and co-founder of HubSpot.swyx [00:00:25]: Co-founder of HubSpot, which I followed for many years, I think 18 years now, gonna be 19 soon. And you caught, you know, people can catch up on your HubSpot story elsewhere. I should also thank Sean Puri, who I've chatted with back and forth, who's been, I guess, getting me in touch with your people. But also, I think like, just giving us a lot of context, because obviously, My First Million joined you guys, and they've been chatting with you guys a lot. So for the business side, we can talk about that, but I kind of wanted to engage your CTO, agent, engineer side of things. So how did you get agent religion?Dharmesh [00:01:00]: Let's see. So I've been working, I'll take like a half step back, a decade or so ago, even though actually more than that. So even before HubSpot, the company I was contemplating that I had named for was called Ingenisoft. And the idea behind Ingenisoft was a natural language interface to business software. Now realize this is 20 years ago, so that was a hard thing to do. But the actual use case that I had in mind was, you know, we had data sitting in business systems like a CRM or something like that. And my kind of what I thought clever at the time. Oh, what if we used email as the kind of interface to get to business software? And the motivation for using email is that it automatically works when you're offline. So imagine I'm getting on a plane or I'm on a plane. There was no internet on planes back then. It's like, oh, I'm going through business cards from an event I went to. I can just type things into an email just to have them all in the backlog. When it reconnects, it sends those emails to a processor that basically kind of parses effectively the commands and updates the software, sends you the file, whatever it is. And there was a handful of commands. I was a little bit ahead of the times in terms of what was actually possible. And I reattempted this natural language thing with a product called ChatSpot that I did back 20...swyx [00:02:12]: Yeah, this is your first post-ChatGPT project.Dharmesh [00:02:14]: I saw it come out. Yeah. And so I've always been kind of fascinated by this natural language interface to software. Because, you know, as software developers, myself included, we've always said, oh, we build intuitive, easy-to-use applications. And it's not intuitive at all, right? Because what we're doing is... We're taking the mental model that's in our head of what we're trying to accomplish with said piece of software and translating that into a series of touches and swipes and clicks and things like that. And there's nothing natural or intuitive about it. And so natural language interfaces, for the first time, you know, whatever the thought is you have in your head and expressed in whatever language that you normally use to talk to yourself in your head, you can just sort of emit that and have software do something. And I thought that was kind of a breakthrough, which it has been. And it's gone. So that's where I first started getting into the journey. I started because now it actually works, right? So once we got ChatGPT and you can take, even with a few-shot example, convert something into structured, even back in the ChatGP 3.5 days, it did a decent job in a few-shot example, convert something to structured text if you knew what kinds of intents you were going to have. And so that happened. And that ultimately became a HubSpot project. But then agents intrigued me because I'm like, okay, well, that's the next step here. So chat's great. Love Chat UX. But if we want to do something even more meaningful, it felt like the next kind of advancement is not this kind of, I'm chatting with some software in a kind of a synchronous back and forth model, is that software is going to do things for me in kind of a multi-step way to try and accomplish some goals. So, yeah, that's when I first got started. It's like, okay, what would that look like? Yeah. And I've been obsessed ever since, by the way.Alessio [00:03:55]: Which goes back to your first experience with it, which is like you're offline. Yeah. And you want to do a task. You don't need to do it right now. You just want to queue it up for somebody to do it for you. Yes. As you think about agents, like, let's start at the easy question, which is like, how do you define an agent? Maybe. You mean the hardest question in the universe? Is that what you mean?Dharmesh [00:04:12]: You said you have an irritating take. I do have an irritating take. I think, well, some number of people have been irritated, including within my own team. So I have a very broad definition for agents, which is it's AI-powered software that accomplishes a goal. Period. That's it. And what irritates people about it is like, well, that's so broad as to be completely non-useful. And I understand that. I understand the criticism. But in my mind, if you kind of fast forward months, I guess, in AI years, the implementation of it, and we're already starting to see this, and we'll talk about this, different kinds of agents, right? So I think in addition to having a usable definition, and I like yours, by the way, and we should talk more about that, that you just came out with, the classification of agents actually is also useful, which is, is it autonomous or non-autonomous? Does it have a deterministic workflow? Does it have a non-deterministic workflow? Is it working synchronously? Is it working asynchronously? Then you have the different kind of interaction modes. Is it a chat agent, kind of like a customer support agent would be? You're having this kind of back and forth. Is it a workflow agent that just does a discrete number of steps? So there's all these different flavors of agents. So if I were to draw it in a Venn diagram, I would draw a big circle that says, this is agents, and then I have a bunch of circles, some overlapping, because they're not mutually exclusive. And so I think that's what's interesting, and we're seeing development along a bunch of different paths, right? So if you look at the first implementation of agent frameworks, you look at Baby AGI and AutoGBT, I think it was, not Autogen, that's the Microsoft one. They were way ahead of their time because they assumed this level of reasoning and execution and planning capability that just did not exist, right? So it was an interesting thought experiment, which is what it was. Even the guy that, I'm an investor in Yohei's fund that did Baby AGI. It wasn't ready, but it was a sign of what was to come. And so the question then is, when is it ready? And so lots of people talk about the state of the art when it comes to agents. I'm a pragmatist, so I think of the state of the practical. It's like, okay, well, what can I actually build that has commercial value or solves actually some discrete problem with some baseline of repeatability or verifiability?swyx [00:06:22]: There was a lot, and very, very interesting. I'm not irritated by it at all. Okay. As you know, I take a... There's a lot of anthropological view or linguistics view. And in linguistics, you don't want to be prescriptive. You want to be descriptive. Yeah. So you're a goals guy. That's the key word in your thing. And other people have other definitions that might involve like delegated trust or non-deterministic work, LLM in the loop, all that stuff. The other thing I was thinking about, just the comment on Baby AGI, LGBT. Yeah. In that piece that you just read, I was able to go through our backlog and just kind of track the winter of agents and then the summer now. Yeah. And it's... We can tell the whole story as an oral history, just following that thread. And it's really just like, I think, I tried to explain the why now, right? Like I had, there's better models, of course. There's better tool use with like, they're just more reliable. Yep. Better tools with MCP and all that stuff. And I'm sure you have opinions on that too. Business model shift, which you like a lot. I just heard you talk about RAS with MFM guys. Yep. Cost is dropping a lot. Yep. Inference is getting faster. There's more model diversity. Yep. Yep. I think it's a subtle point. It means that like, you have different models with different perspectives. You don't get stuck in the basin of performance of a single model. Sure. You can just get out of it by just switching models. Yep. Multi-agent research and RL fine tuning. So I just wanted to let you respond to like any of that.Dharmesh [00:07:44]: Yeah. A couple of things. Connecting the dots on the kind of the definition side of it. So we'll get the irritation out of the way completely. I have one more, even more irritating leap on the agent definition thing. So here's the way I think about it. By the way, the kind of word agent, I looked it up, like the English dictionary definition. The old school agent, yeah. Is when you have someone or something that does something on your behalf, like a travel agent or a real estate agent acts on your behalf. It's like proxy, which is a nice kind of general definition. So the other direction I'm sort of headed, and it's going to tie back to tool calling and MCP and things like that, is if you, and I'm not a biologist by any stretch of the imagination, but we have these single-celled organisms, right? Like the simplest possible form of what one would call life. But it's still life. It just happens to be single-celled. And then you can combine cells and then cells become specialized over time. And you have much more sophisticated organisms, you know, kind of further down the spectrum. In my mind, at the most fundamental level, you can almost think of having atomic agents. What is the simplest possible thing that's an agent that can still be called an agent? What is the equivalent of a kind of single-celled organism? And the reason I think that's useful is right now we're headed down the road, which I think is very exciting around tool use, right? That says, okay, the LLMs now can be provided a set of tools that it calls to accomplish whatever it needs to accomplish in the kind of furtherance of whatever goal it's trying to get done. And I'm not overly bothered by it, but if you think about it, if you just squint a little bit and say, well, what if everything was an agent? And what if tools were actually just atomic agents? Because then it's turtles all the way down, right? Then it's like, oh, well, all that's really happening with tool use is that we have a network of agents that know about each other through something like an MMCP and can kind of decompose a particular problem and say, oh, I'm going to delegate this to this set of agents. And why do we need to draw this distinction between tools, which are functions most of the time? And an actual agent. And so I'm going to write this irritating LinkedIn post, you know, proposing this. It's like, okay. And I'm not suggesting we should call even functions, you know, call them agents. But there is a certain amount of elegance that happens when you say, oh, we can just reduce it down to one primitive, which is an agent that you can combine in complicated ways to kind of raise the level of abstraction and accomplish higher order goals. Anyway, that's my answer. I'd say that's a success. Thank you for coming to my TED Talk on agent definitions.Alessio [00:09:54]: How do you define the minimum viable agent? Do you already have a definition for, like, where you draw the line between a cell and an atom? Yeah.Dharmesh [00:10:02]: So in my mind, it has to, at some level, use AI in order for it to—otherwise, it's just software. It's like, you know, we don't need another word for that. And so that's probably where I draw the line. So then the question, you know, the counterargument would be, well, if that's true, then lots of tools themselves are actually not agents because they're just doing a database call or a REST API call or whatever it is they're doing. And that does not necessarily qualify them, which is a fair counterargument. And I accept that. It's like a good argument. I still like to think about—because we'll talk about multi-agent systems, because I think—so we've accepted, which I think is true, lots of people have said it, and you've hopefully combined some of those clips of really smart people saying this is the year of agents, and I completely agree, it is the year of agents. But then shortly after that, it's going to be the year of multi-agent systems or multi-agent networks. I think that's where it's going to be headed next year. Yeah.swyx [00:10:54]: Opening eyes already on that. Yeah. My quick philosophical engagement with you on this. I often think about kind of the other spectrum, the other end of the cell spectrum. So single cell is life, multi-cell is life, and you clump a bunch of cells together in a more complex organism, they become organs, like an eye and a liver or whatever. And then obviously we consider ourselves one life form. There's not like a lot of lives within me. I'm just one life. And now, obviously, I don't think people don't really like to anthropomorphize agents and AI. Yeah. But we are extending our consciousness and our brain and our functionality out into machines. I just saw you were a Bee. Yeah. Which is, you know, it's nice. I have a limitless pendant in my pocket.Dharmesh [00:11:37]: I got one of these boys. Yeah.swyx [00:11:39]: I'm testing it all out. You know, got to be early adopters. But like, we want to extend our personal memory into these things so that we can be good at the things that we're good at. And, you know, machines are good at it. Machines are there. So like, my definition of life is kind of like going outside of my own body now. I don't know if you've ever had like reflections on that. Like how yours. How our self is like actually being distributed outside of you. Yeah.Dharmesh [00:12:01]: I don't fancy myself a philosopher. But you went there. So yeah, I did go there. I'm fascinated by kind of graphs and graph theory and networks and have been for a long, long time. And to me, we're sort of all nodes in this kind of larger thing. It just so happens that we're looking at individual kind of life forms as they exist right now. But so the idea is when you put a podcast out there, there's these little kind of nodes you're putting out there of like, you know, conceptual ideas. Once again, you have varying kind of forms of those little nodes that are up there and are connected in varying and sundry ways. And so I just think of myself as being a node in a massive, massive network. And I'm producing more nodes as I put content or ideas. And, you know, you spend some portion of your life collecting dots, experiences, people, and some portion of your life then connecting dots from the ones that you've collected over time. And I found that really interesting things happen and you really can't know in advance how those dots are necessarily going to connect in the future. And that's, yeah. So that's my philosophical take. That's the, yes, exactly. Coming back.Alessio [00:13:04]: Yep. Do you like graph as an agent? Abstraction? That's been one of the hot topics with LandGraph and Pydantic and all that.Dharmesh [00:13:11]: I do. The thing I'm more interested in terms of use of graphs, and there's lots of work happening on that now, is graph data stores as an alternative in terms of knowledge stores and knowledge graphs. Yeah. Because, you know, so I've been in software now 30 plus years, right? So it's not 10,000 hours. It's like 100,000 hours that I've spent doing this stuff. And so I've grew up with, so back in the day, you know, I started on mainframes. There was a product called IMS from IBM, which is basically an index database, what we'd call like a key value store today. Then we've had relational databases, right? We have tables and columns and foreign key relationships. We all know that. We have document databases like MongoDB, which is sort of a nested structure keyed by a specific index. We have vector stores, vector embedding database. And graphs are interesting for a couple of reasons. One is, so it's not classically structured in a relational way. When you say structured database, to most people, they're thinking tables and columns and in relational database and set theory and all that. Graphs still have structure, but it's not the tables and columns structure. And you could wonder, and people have made this case, that they are a better representation of knowledge for LLMs and for AI generally than other things. So that's kind of thing number one conceptually, and that might be true, I think is possibly true. And the other thing that I really like about that in the context of, you know, I've been in the context of data stores for RAG is, you know, RAG, you say, oh, I have a million documents, I'm going to build the vector embeddings, I'm going to come back with the top X based on the semantic match, and that's fine. All that's very, very useful. But the reality is something gets lost in the chunking process and the, okay, well, those tend, you know, like, you don't really get the whole picture, so to speak, and maybe not even the right set of dimensions on the kind of broader picture. And it makes intuitive sense to me that if we did capture it properly in a graph form, that maybe that feeding into a RAG pipeline will actually yield better results for some use cases, I don't know, but yeah.Alessio [00:15:03]: And do you feel like at the core of it, there's this difference between imperative and declarative programs? Because if you think about HubSpot, it's like, you know, people and graph kind of goes hand in hand, you know, but I think maybe the software before was more like primary foreign key based relationship, versus now the models can traverse through the graph more easily.Dharmesh [00:15:22]: Yes. So I like that representation. There's something. It's just conceptually elegant about graphs and just from the representation of it, they're much more discoverable, you can kind of see it, there's observability to it, versus kind of embeddings, which you can't really do much with as a human. You know, once they're in there, you can't pull stuff back out. But yeah, I like that kind of idea of it. And the other thing that's kind of, because I love graphs, I've been long obsessed with PageRank from back in the early days. And, you know, one of the kind of simplest algorithms in terms of coming up, you know, with a phone, everyone's been exposed to PageRank. And the idea is that, and so I had this other idea for a project, not a company, and I have hundreds of these, called NodeRank, is to be able to take the idea of PageRank and apply it to an arbitrary graph that says, okay, I'm going to define what authority looks like and say, okay, well, that's interesting to me, because then if you say, I'm going to take my knowledge store, and maybe this person that contributed some number of chunks to the graph data store has more authority on this particular use case or prompt that's being submitted than this other one that may, or maybe this one was more. popular, or maybe this one has, whatever it is, there should be a way for us to kind of rank nodes in a graph and sort them in some, some useful way. Yeah.swyx [00:16:34]: So I think that's generally useful for, for anything. I think the, the problem, like, so even though at my conferences, GraphRag is super popular and people are getting knowledge, graph religion, and I will say like, it's getting space, getting traction in two areas, conversation memory, and then also just rag in general, like the, the, the document data. Yeah. It's like a source. Most ML practitioners would say that knowledge graph is kind of like a dirty word. The graph database, people get graph religion, everything's a graph, and then they, they go really hard into it and then they get a, they get a graph that is too complex to navigate. Yes. And so like the, the, the simple way to put it is like you at running HubSpot, you know, the power of graphs, the way that Google has pitched them for many years, but I don't suspect that HubSpot itself uses a knowledge graph. No. Yeah.Dharmesh [00:17:26]: So when is it over engineering? Basically? It's a great question. I don't know. So the question now, like in AI land, right, is the, do we necessarily need to understand? So right now, LLMs for, for the most part are somewhat black boxes, right? We sort of understand how the, you know, the algorithm itself works, but we really don't know what's going on in there and, and how things come out. So if a graph data store is able to produce the outcomes we want, it's like, here's a set of queries I want to be able to submit and then it comes out with useful content. Maybe the underlying data store is as opaque as a vector embeddings or something like that, but maybe it's fine. Maybe we don't necessarily need to understand it to get utility out of it. And so maybe if it's messy, that's okay. Um, that's, it's just another form of lossy compression. Uh, it's just lossy in a way that we just don't completely understand in terms of, because it's going to grow organically. Uh, and it's not structured. It's like, ah, we're just gonna throw a bunch of stuff in there. Let the, the equivalent of the embedding algorithm, whatever they called in graph land. Um, so the one with the best results wins. I think so. Yeah.swyx [00:18:26]: Or is this the practical side of me is like, yeah, it's, if it's useful, we don't necessarilyDharmesh [00:18:30]: need to understand it.swyx [00:18:30]: I have, I mean, I'm happy to push back as long as you want. Uh, it's not practical to evaluate like the 10 different options out there because it takes time. It takes people, it takes, you know, resources, right? Set. That's the first thing. Second thing is your evals are typically on small things and some things only work at scale. Yup. Like graphs. Yup.Dharmesh [00:18:46]: Yup. That's, yeah, no, that's fair. And I think this is one of the challenges in terms of implementation of graph databases is that the most common approach that I've seen developers do, I've done it myself, is that, oh, I've got a Postgres database or a MySQL or whatever. I can represent a graph with a very set of tables with a parent child thing or whatever. And that sort of gives me the ability, uh, why would I need anything more than that? And the answer is, well, if you don't need anything more than that, you don't need anything more than that. But there's a high chance that you're sort of missing out on the actual value that, uh, the graph representation gives you. Which is the ability to traverse the graph, uh, efficiently in ways that kind of going through the, uh, traversal in a relational database form, even though structurally you have the data, practically you're not gonna be able to pull it out in, in useful ways. Uh, so you wouldn't like represent a social graph, uh, in, in using that kind of relational table model. It just wouldn't scale. It wouldn't work.swyx [00:19:36]: Uh, yeah. Uh, I think we want to move on to MCP. Yeah. But I just want to, like, just engineering advice. Yeah. Uh, obviously you've, you've, you've run, uh, you've, you've had to do a lot of projects and run a lot of teams. Do you have a general rule for over-engineering or, you know, engineering ahead of time? You know, like, because people, we know premature engineering is the root of all evil. Yep. But also sometimes you just have to. Yep. When do you do it? Yes.Dharmesh [00:19:59]: It's a great question. This is, uh, a question as old as time almost, which is what's the right and wrong levels of abstraction. That's effectively what, uh, we're answering when we're trying to do engineering. I tend to be a pragmatist, right? So here's the thing. Um, lots of times doing something the right way. Yeah. It's like a marginal increased cost in those cases. Just do it the right way. And this is what makes a, uh, a great engineer or a good engineer better than, uh, a not so great one. It's like, okay, all things being equal. If it's going to take you, you know, roughly close to constant time anyway, might as well do it the right way. Like, so do things well, then the question is, okay, well, am I building a framework as the reusable library? To what degree, uh, what am I anticipating in terms of what's going to need to change in this thing? Uh, you know, along what dimension? And then I think like a business person in some ways, like what's the return on calories, right? So, uh, and you look at, um, energy, the expected value of it's like, okay, here are the five possible things that could happen, uh, try to assign probabilities like, okay, well, if there's a 50% chance that we're going to go down this particular path at some day, like, or one of these five things is going to happen and it costs you 10% more to engineer for that. It's basically, it's something that yields a kind of interest compounding value. Um, as you get closer to the time of, of needing that versus having to take on debt, which is when you under engineer it, you're taking on debt. You're going to have to pay off when you do get to that eventuality where something happens. One thing as a pragmatist, uh, so I would rather under engineer something than over engineer it. If I were going to err on the side of something, and here's the reason is that when you under engineer it, uh, yes, you take on tech debt, uh, but the interest rate is relatively known and payoff is very, very possible, right? Which is, oh, I took a shortcut here as a result of which now this thing that should have taken me a week is now going to take me four weeks. Fine. But if that particular thing that you thought might happen, never actually, you never have that use case transpire or just doesn't, it's like, well, you just save yourself time, right? And that has value because you were able to do other things instead of, uh, kind of slightly over-engineering it away, over-engineering it. But there's no perfect answers in art form in terms of, uh, and yeah, we'll, we'll bring kind of this layers of abstraction back on the code generation conversation, which we'll, uh, I think I have later on, butAlessio [00:22:05]: I was going to ask, we can just jump ahead quickly. Yeah. Like, as you think about vibe coding and all that, how does the. Yeah. Percentage of potential usefulness change when I feel like we over-engineering a lot of times it's like the investment in syntax, it's less about the investment in like arc exacting. Yep. Yeah. How does that change your calculus?Dharmesh [00:22:22]: A couple of things, right? One is, um, so, you know, going back to that kind of ROI or a return on calories, kind of calculus or heuristic you think through, it's like, okay, well, what is it going to cost me to put this layer of abstraction above the code that I'm writing now, uh, in anticipating kind of future needs. If the cost of fixing, uh, or doing under engineering right now. Uh, we'll trend towards zero that says, okay, well, I don't have to get it right right now because even if I get it wrong, I'll run the thing for six hours instead of 60 minutes or whatever. It doesn't really matter, right? Like, because that's going to trend towards zero to be able, the ability to refactor a code. Um, and because we're going to not that long from now, we're going to have, you know, large code bases be able to exist, uh, you know, as, as context, uh, for a code generation or a code refactoring, uh, model. So I think it's going to make it, uh, make the case for under engineering, uh, even stronger. Which is why I take on that cost. You just pay the interest when you get there, it's not, um, just go on with your life vibe coded and, uh, come back when you need to. Yeah.Alessio [00:23:18]: Sometimes I feel like there's no decision-making in some things like, uh, today I built a autosave for like our internal notes platform and I literally just ask them cursor. Can you add autosave? Yeah. I don't know if it's over under engineer. Yep. I just vibe coded it. Yep. And I feel like at some point we're going to get to the point where the models kindDharmesh [00:23:36]: of decide where the right line is, but this is where the, like the, in my mind, the danger is, right? So there's two sides to this. One is the cost of kind of development and coding and things like that stuff that, you know, we talk about. But then like in your example, you know, one of the risks that we have is that because adding a feature, uh, like a save or whatever the feature might be to a product as that price tends towards zero, are we going to be less discriminant about what features we add as a result of making more product products more complicated, which has a negative impact on the user and navigate negative impact on the business. Um, and so that's the thing I worry about if it starts to become too easy, are we going to be. Too promiscuous in our, uh, kind of extension, adding product extensions and things like that. It's like, ah, why not add X, Y, Z or whatever back then it was like, oh, we only have so many engineering hours or story points or however you measure things. Uh, that least kept us in check a little bit. Yeah.Alessio [00:24:22]: And then over engineering, you're like, yeah, it's kind of like you're putting that on yourself. Yeah. Like now it's like the models don't understand that if they add too much complexity, it's going to come back to bite them later. Yep. So they just do whatever they want to do. Yeah. And I'm curious where in the workflow that's going to be, where it's like, Hey, this is like the amount of complexity and over-engineering you can do before you got to ask me if we should actually do it versus like do something else.Dharmesh [00:24:45]: So you know, we've already, let's like, we're leaving this, uh, in the code generation world, this kind of compressed, um, cycle time. Right. It's like, okay, we went from auto-complete, uh, in the GitHub co-pilot to like, oh, finish this particular thing and hit tab to a, oh, I sort of know your file or whatever. I can write out a full function to you to now I can like hold a bunch of the context in my head. Uh, so we can do app generation, which we have now with lovable and bolt and repletage. Yeah. Association and other things. So then the question is, okay, well, where does it naturally go from here? So we're going to generate products. Make sense. We might be able to generate platforms as though I want a platform for ERP that does this, whatever. And that includes the API's includes the product and the UI, and all the things that make for a platform. There's no nothing that says we would stop like, okay, can you generate an entire software company someday? Right. Uh, with the platform and the monetization and the go-to-market and the whatever. And you know, that that's interesting to me in terms of, uh, you know, what, when you take it to almost ludicrous levels. of abstract.swyx [00:25:39]: It's like, okay, turn it to 11. You mentioned vibe coding, so I have to, this is a blog post I haven't written, but I'm kind of exploring it. Is the junior engineer dead?Dharmesh [00:25:49]: I don't think so. I think what will happen is that the junior engineer will be able to, if all they're bringing to the table is the fact that they are a junior engineer, then yes, they're likely dead. But hopefully if they can communicate with carbon-based life forms, they can interact with product, if they're willing to talk to customers, they can take their kind of basic understanding of engineering and how kind of software works. I think that has value. So I have a 14-year-old right now who's taking Python programming class, and some people ask me, it's like, why is he learning coding? And my answer is, is because it's not about the syntax, it's not about the coding. What he's learning is like the fundamental thing of like how things work. And there's value in that. I think there's going to be timeless value in systems thinking and abstractions and what that means. And whether functions manifested as math, which he's going to get exposed to regardless, or there are some core primitives to the universe, I think, that the more you understand them, those are what I would kind of think of as like really large dots in your life that will have a higher gravitational pull and value to them that you'll then be able to. So I want him to collect those dots, and he's not resisting. So it's like, okay, while he's still listening to me, I'm going to have him do things that I think will be useful.swyx [00:26:59]: You know, part of one of the pitches that I evaluated for AI engineer is a term. And the term is that maybe the traditional interview path or career path of software engineer goes away, which is because what's the point of lead code? Yeah. And, you know, it actually matters more that you know how to work with AI and to implement the things that you want. Yep.Dharmesh [00:27:16]: That's one of the like interesting things that's happened with generative AI. You know, you go from machine learning and the models and just that underlying form, which is like true engineering, right? Like the actual, what I call real engineering. I don't think of myself as a real engineer, actually. I'm a developer. But now with generative AI. We call it AI and it's obviously got its roots in machine learning, but it just feels like fundamentally different to me. Like you have the vibe. It's like, okay, well, this is just a whole different approach to software development to so many different things. And so I'm wondering now, it's like an AI engineer is like, if you were like to draw the Venn diagram, it's interesting because the cross between like AI things, generative AI and what the tools are capable of, what the models do, and this whole new kind of body of knowledge that we're still building out, it's still very young, intersected with kind of classic engineering, software engineering. Yeah.swyx [00:28:04]: I just described the overlap as it separates out eventually until it's its own thing, but it's starting out as a software. Yeah.Alessio [00:28:11]: That makes sense. So to close the vibe coding loop, the other big hype now is MCPs. Obviously, I would say Cloud Desktop and Cursor are like the two main drivers of MCP usage. I would say my favorite is the Sentry MCP. I can pull in errors and then you can just put the context in Cursor. How do you think about that abstraction layer? Does it feel... Does it feel almost too magical in a way? Do you think it's like you get enough? Because you don't really see how the server itself is then kind of like repackaging theDharmesh [00:28:41]: information for you? I think MCP as a standard is one of the better things that's happened in the world of AI because a standard needed to exist and absent a standard, there was a set of things that just weren't possible. Now, we can argue whether it's the best possible manifestation of a standard or not. Does it do too much? Does it do too little? I get that, but it's just simple enough to both be useful and unobtrusive. It's understandable and adoptable by mere mortals, right? It's not overly complicated. You know, a reasonable engineer can put a stand up an MCP server relatively easily. The thing that has me excited about it is like, so I'm a big believer in multi-agent systems. And so that's going back to our kind of this idea of an atomic agent. So imagine the MCP server, like obviously it calls tools, but the way I think about it, so I'm working on my current passion project is agent.ai. And we'll talk more about that in a little bit. More about the, I think we should, because I think it's interesting not to promote the project at all, but there's some interesting ideas in there. One of which is around, we're going to need a mechanism for, if agents are going to collaborate and be able to delegate, there's going to need to be some form of discovery and we're going to need some standard way. It's like, okay, well, I just need to know what this thing over here is capable of. We're going to need a registry, which Anthropic's working on. I'm sure others will and have been doing directories of, and there's going to be a standard around that too. How do you build out a directory of MCP servers? I think that's going to unlock so many things just because, and we're already starting to see it. So I think MCP or something like it is going to be the next major unlock because it allows systems that don't know about each other, don't need to, it's that kind of decoupling of like Sentry and whatever tools someone else was building. And it's not just about, you know, Cloud Desktop or things like, even on the client side, I think we're going to see very interesting consumers of MCP, MCP clients versus just the chat body kind of things. Like, you know, Cloud Desktop and Cursor and things like that. But yeah, I'm very excited about MCP in that general direction.swyx [00:30:39]: I think the typical cynical developer take, it's like, we have OpenAPI. Yeah. What's the new thing? I don't know if you have a, do you have a quick MCP versus everything else? Yeah.Dharmesh [00:30:49]: So it's, so I like OpenAPI, right? So just a descriptive thing. It's OpenAPI. OpenAPI. Yes, that's what I meant. So it's basically a self-documenting thing. We can do machine-generated, lots of things from that output. It's a structured definition of an API. I get that, love it. But MCPs sort of are kind of use case specific. They're perfect for exactly what we're trying to use them for around LLMs in terms of discovery. It's like, okay, I don't necessarily need to know kind of all this detail. And so right now we have, we'll talk more about like MCP server implementations, but We will? I think, I don't know. Maybe we won't. At least it's in my head. It's like a back processor. But I do think MCP adds value above OpenAPI. It's, yeah, just because it solves this particular thing. And if we had come to the world, which we have, like, it's like, hey, we already have OpenAPI. It's like, if that were good enough for the universe, the universe would have adopted it already. There's a reason why MCP is taking office because marginally adds something that was missing before and doesn't go too far. And so that's why the kind of rate of adoption, you folks have written about this and talked about it. Yeah, why MCP won. Yeah. And it won because the universe decided that this was useful and maybe it gets supplanted by something else. Yeah. And maybe we discover, oh, maybe OpenAPI was good enough the whole time. I doubt that.swyx [00:32:09]: The meta lesson, this is, I mean, he's an investor in DevTools companies. I work in developer experience at DevRel in DevTools companies. Yep. Everyone wants to own the standard. Yeah. I'm sure you guys have tried to launch your own standards. Actually, it's Houseplant known for a standard, you know, obviously inbound marketing. But is there a standard or protocol that you ever tried to push? No.Dharmesh [00:32:30]: And there's a reason for this. Yeah. Is that? And I don't mean, need to mean, speak for the people of HubSpot, but I personally. You kind of do. I'm not smart enough. That's not the, like, I think I have a. You're smart. Not enough for that. I'm much better off understanding the standards that are out there. And I'm more on the composability side. Let's, like, take the pieces of technology that exist out there, combine them in creative, unique ways. And I like to consume standards. I don't like to, and that's not that I don't like to create them. I just don't think I have the, both the raw wattage or the credibility. It's like, okay, well, who the heck is Dharmesh, and why should we adopt a standard he created?swyx [00:33:07]: Yeah, I mean, there are people who don't monetize standards, like OpenTelemetry is a big standard, and LightStep never capitalized on that.Dharmesh [00:33:15]: So, okay, so if I were to do a standard, there's two things that have been in my head in the past. I was one around, a very, very basic one around, I don't even have the domain, I have a domain for everything, for open marketing. Because the issue we had in HubSpot grew up in the marketing space. There we go. There was no standard around data formats and things like that. It doesn't go anywhere. But the other one, and I did not mean to go here, but I'm going to go here. It's called OpenGraph. I know the term was already taken, but it hasn't been used for like 15 years now for its original purpose. But what I think should exist in the world is right now, our information, all of us, nodes are in the social graph at Meta or the professional graph at LinkedIn. Both of which are actually relatively closed in actually very annoying ways. Like very, very closed, right? Especially LinkedIn. Especially LinkedIn. I personally believe that if it's my data, and if I would get utility out of it being open, I should be able to make my data open or publish it in whatever forms that I choose, as long as I have control over it as opt-in. So the idea is around OpenGraph that says, here's a standard, here's a way to publish it. I should be able to go to OpenGraph.org slash Dharmesh dot JSON and get it back. And it's like, here's your stuff, right? And I can choose along the way and people can write to it and I can prove. And there can be an entire system. And if I were to do that, I would do it as a... Like a public benefit, non-profit-y kind of thing, as this is a contribution to society. I wouldn't try to commercialize that. Have you looked at AdProto? What's that? AdProto.swyx [00:34:43]: It's the protocol behind Blue Sky. Okay. My good friend, Dan Abramov, who was the face of React for many, many years, now works there. And he actually did a talk that I can send you, which basically kind of tries to articulate what you just said. But he does, he loves doing these like really great analogies, which I think you'll like. Like, you know, a lot of our data is behind a handle, behind a domain. Yep. So he's like, all right, what if we flip that? What if it was like our handle and then the domain? Yep. So, and that's really like your data should belong to you. Yep. And I should not have to wait 30 days for my Twitter data to export. Yep.Dharmesh [00:35:19]: you should be able to at least be able to automate it or do like, yes, I should be able to plug it into an agentic thing. Yeah. Yes. I think we're... Because so much of our data is... Locked up. I think the trick here isn't that standard. It is getting the normies to care.swyx [00:35:37]: Yeah. Because normies don't care.Dharmesh [00:35:38]: That's true. But building on that, normies don't care. So, you know, privacy is a really hot topic and an easy word to use, but it's not a binary thing. Like there are use cases where, and we make these choices all the time, that I will trade, not all privacy, but I will trade some privacy for some productivity gain or some benefit to me that says, oh, I don't care about that particular data being online if it gives me this in return, or I don't mind sharing this information with this company.Alessio [00:36:02]: If I'm getting, you know, this in return, but that sort of should be my option. I think now with computer use, you can actually automate some of the exports. Yes. Like something we've been doing internally is like everybody exports their LinkedIn connections. Yep. And then internally, we kind of merge them together to see how we can connect our companies to customers or things like that.Dharmesh [00:36:21]: And not to pick on LinkedIn, but since we're talking about it, but they feel strongly enough on the, you know, do not take LinkedIn data that they will block even browser use kind of things or whatever. They go to great, great lengths, even to see patterns of usage. And it says, oh, there's no way you could have, you know, gotten that particular thing or whatever without, and it's, so it's, there's...swyx [00:36:42]: Wasn't there a Supreme Court case that they lost? Yeah.Dharmesh [00:36:45]: So the one they lost was around someone that was scraping public data that was on the public internet. And that particular company had not signed any terms of service or whatever. It's like, oh, I'm just taking data that's on, there was no, and so that's why they won. But now, you know, the question is around, can LinkedIn... I think they can. Like, when you use, as a user, you use LinkedIn, you are signing up for their terms of service. And if they say, well, this kind of use of your LinkedIn account that violates our terms of service, they can shut your account down, right? They can. And they, yeah, so, you know, we don't need to make this a discussion. By the way, I love the company, don't get me wrong. I'm an avid user of the product. You know, I've got... Yeah, I mean, you've got over a million followers on LinkedIn, I think. Yeah, I do. And I've known people there for a long, long time, right? And I have lots of respect. And I understand even where the mindset originally came from of this kind of members-first approach to, you know, a privacy-first. I sort of get that. But sometimes you sort of have to wonder, it's like, okay, well, that was 15, 20 years ago. There's likely some controlled ways to expose some data on some member's behalf and not just completely be a binary. It's like, no, thou shalt not have the data.swyx [00:37:54]: Well, just pay for sales navigator.Alessio [00:37:57]: Before we move to the next layer of instruction, anything else on MCP you mentioned? Let's move back and then I'll tie it back to MCPs.Dharmesh [00:38:05]: So I think the... Open this with agent. Okay, so I'll start with... Here's my kind of running thesis, is that as AI and agents evolve, which they're doing very, very quickly, we're going to look at them more and more. I don't like to anthropomorphize. We'll talk about why this is not that. Less as just like raw tools and more like teammates. They'll still be software. They should self-disclose as being software. I'm totally cool with that. But I think what's going to happen is that in the same way you might collaborate with a team member on Slack or Teams or whatever you use, you can imagine a series of agents that do specific things just like a team member might do, that you can delegate things to. You can collaborate. You can say, hey, can you take a look at this? Can you proofread that? Can you try this? You can... Whatever it happens to be. So I think it is... I will go so far as to say it's inevitable that we're going to have hybrid teams someday. And what I mean by hybrid teams... So back in the day, hybrid teams were, oh, well, you have some full-time employees and some contractors. Then it was like hybrid teams are some people that are in the office and some that are remote. That's the kind of form of hybrid. The next form of hybrid is like the carbon-based life forms and agents and AI and some form of software. So let's say we temporarily stipulate that I'm right about that over some time horizon that eventually we're going to have these kind of digitally hybrid teams. So if that's true, then the question you sort of ask yourself is that then what needs to exist in order for us to get the full value of that new model? It's like, okay, well... You sort of need to... It's like, okay, well, how do I... If I'm building a digital team, like, how do I... Just in the same way, if I'm interviewing for an engineer or a designer or a PM, whatever, it's like, well, that's why we have professional networks, right? It's like, oh, they have a presence on likely LinkedIn. I can go through that semi-structured, structured form, and I can see the experience of whatever, you know, self-disclosed. But, okay, well, agents are going to need that someday. And so I'm like, okay, well, this seems like a thread that's worth pulling on. That says, okay. So I... So agent.ai is out there. And it's LinkedIn for agents. It's LinkedIn for agents. It's a professional network for agents. And the more I pull on that thread, it's like, okay, well, if that's true, like, what happens, right? It's like, oh, well, they have a profile just like anyone else, just like a human would. It's going to be a graph underneath, just like a professional network would be. It's just that... And you can have its, you know, connections and follows, and agents should be able to post. That's maybe how they do release notes. Like, oh, I have this new version. Whatever they decide to post, it should just be able to... Behave as a node on the network of a professional network. As it turns out, the more I think about that and pull on that thread, the more and more things, like, start to make sense to me. So it may be more than just a pure professional network. So my original thought was, okay, well, it's a professional network and agents as they exist out there, which I think there's going to be more and more of, will kind of exist on this network and have the profile. But then, and this is always dangerous, I'm like, okay, I want to see a world where thousands of agents are out there in order for the... Because those digital employees, the digital workers don't exist yet in any meaningful way. And so then I'm like, oh, can I make that easier for, like... And so I have, as one does, it's like, oh, I'll build a low-code platform for building agents. How hard could that be, right? Like, very hard, as it turns out. But it's been fun. So now, agent.ai has 1.3 million users. 3,000 people have actually, you know, built some variation of an agent, sometimes just for their own personal productivity. About 1,000 of which have been published. And the reason this comes back to MCP for me, so imagine that and other networks, since I know agent.ai. So right now, we have an MCP server for agent.ai that exposes all the internally built agents that we have that do, like, super useful things. Like, you know, I have access to a Twitter API that I can subsidize the cost. And I can say, you know, if you're looking to build something for social media, these kinds of things, with a single API key, and it's all completely free right now, I'm funding it. That's a useful way for it to work. And then we have a developer to say, oh, I have this idea. I don't have to worry about open AI. I don't have to worry about, now, you know, this particular model is better. It has access to all the models with one key. And we proxy it kind of behind the scenes. And then expose it. So then we get this kind of community effect, right? That says, oh, well, someone else may have built an agent to do X. Like, I have an agent right now that I built for myself to do domain valuation for website domains because I'm obsessed with domains, right? And, like, there's no efficient market for domains. There's no Zillow for domains right now that tells you, oh, here are what houses in your neighborhood sold for. It's like, well, why doesn't that exist? We should be able to solve that problem. And, yes, you're still guessing. Fine. There should be some simple heuristic. So I built that. It's like, okay, well, let me go look for past transactions. You say, okay, I'm going to type in agent.ai, agent.com, whatever domain. What's it actually worth? I'm looking at buying it. It can go and say, oh, which is what it does. It's like, I'm going to go look at are there any published domain transactions recently that are similar, either use the same word, same top-level domain, whatever it is. And it comes back with an approximate value, and it comes back with its kind of rationale for why it picked the value and comparable transactions. Oh, by the way, this domain sold for published. Okay. So that agent now, let's say, existed on the web, on agent.ai. Then imagine someone else says, oh, you know, I want to build a brand-building agent for startups and entrepreneurs to come up with names for their startup. Like a common problem, every startup is like, ah, I don't know what to call it. And so they type in five random words that kind of define whatever their startup is. And you can do all manner of things, one of which is like, oh, well, I need to find the domain for it. What are possible choices? Now it's like, okay, well, it would be nice to know if there's an aftermarket price for it, if it's listed for sale. Awesome. Then imagine calling this valuation agent. It's like, okay, well, I want to find where the arbitrage is, where the agent valuation tool says this thing is worth $25,000. It's listed on GoDaddy for $5,000. It's close enough. Let's go do that. Right? And that's a kind of composition use case that in my future state. Thousands of agents on the network, all discoverable through something like MCP. And then you as a developer of agents have access to all these kind of Lego building blocks based on what you're trying to solve. Then you blend in orchestration, which is getting better and better with the reasoning models now. Just describe the problem that you have. Now, the next layer that we're all contending with is that how many tools can you actually give an LLM before the LLM breaks? That number used to be like 15 or 20 before you kind of started to vary dramatically. And so that's the thing I'm thinking about now. It's like, okay, if I want to... If I want to expose 1,000 of these agents to a given LLM, obviously I can't give it all 1,000. Is there some intermediate layer that says, based on your prompt, I'm going to make a best guess at which agents might be able to be helpful for this particular thing? Yeah.Alessio [00:44:37]: Yeah, like RAG for tools. Yep. I did build the Latent Space Researcher on agent.ai. Okay. Nice. Yeah, that seems like, you know, then there's going to be a Latent Space Scheduler. And then once I schedule a research, you know, and you build all of these things. By the way, my apologies for the user experience. You realize I'm an engineer. It's pretty good.swyx [00:44:56]: I think it's a normie-friendly thing. Yeah. That's your magic. HubSpot does the same thing.Alessio [00:45:01]: Yeah, just to like quickly run through it. You can basically create all these different steps. And these steps are like, you know, static versus like variable-driven things. How did you decide between this kind of like low-code-ish versus doing, you know, low-code with code backend versus like not exposing that at all? Any fun design decisions? Yeah. And this is, I think...Dharmesh [00:45:22]: I think lots of people are likely sitting in exactly my position right now, coming through the choosing between deterministic. Like if you're like in a business or building, you know, some sort of agentic thing, do you decide to do a deterministic thing? Or do you go non-deterministic and just let the alum handle it, right, with the reasoning models? The original idea and the reason I took the low-code stepwise, a very deterministic approach. A, the reasoning models did not exist at that time. That's thing number one. Thing number two is if you can get... If you know in your head... If you know in your head what the actual steps are to accomplish whatever goal, why would you leave that to chance? There's no upside. There's literally no upside. Just tell me, like, what steps do you need executed? So right now what I'm playing with... So one thing we haven't talked about yet, and people don't talk about UI and agents. Right now, the primary interaction model... Or they don't talk enough about it. I know some people have. But it's like, okay, so we're used to the chatbot back and forth. Fine. I get that. But I think we're going to move to a blend of... Some of those things are going to be synchronous as they are now. But some are going to be... Some are going to be async. It's just going to put it in a queue, just like... And this goes back to my... Man, I talk fast. But I have this... I only have one other speed. It's even faster. So imagine it's like if you're working... So back to my, oh, we're going to have these hybrid digital teams. Like, you would not go to a co-worker and say, I'm going to ask you to do this thing, and then sit there and wait for them to go do it. Like, that's not how the world works. So it's nice to be able to just, like, hand something off to someone. It's like, okay, well, maybe I expect a response in an hour or a day or something like that.Dharmesh [00:46:52]: In terms of when things need to happen. So the UI around agents. So if you look at the output of agent.ai agents right now, they are the simplest possible manifestation of a UI, right? That says, oh, we have inputs of, like, four different types. Like, we've got a dropdown, we've got multi-select, all the things. It's like back in HTML, the original HTML 1.0 days, right? Like, you're the smallest possible set of primitives for a UI. And it just says, okay, because we need to collect some information from the user, and then we go do steps and do things. And generate some output in HTML or markup are the two primary examples. So the thing I've been asking myself, if I keep going down that path. So people ask me, I get requests all the time. It's like, oh, can you make the UI sort of boring? I need to be able to do this, right? And if I keep pulling on that, it's like, okay, well, now I've built an entire UI builder thing. Where does this end? And so I think the right answer, and this is what I'm going to be backcoding once I get done here, is around injecting a code generation UI generation into, the agent.ai flow, right? As a builder, you're like, okay, I'm going to describe the thing that I want, much like you would do in a vibe coding world. But instead of generating the entire app, it's going to generate the UI that exists at some point in either that deterministic flow or something like that. It says, oh, here's the thing I'm trying to do. Go generate the UI for me. And I can go through some iterations. And what I think of it as a, so it's like, I'm going to generate the code, generate the code, tweak it, go through this kind of prompt style, like we do with vibe coding now. And at some point, I'm going to be happy with it. And I'm going to hit save. And that's going to become the action in that particular step. It's like a caching of the generated code that I can then, like incur any inference time costs. It's just the actual code at that point.Alessio [00:48:29]: Yeah, I invested in a company called E2B, which does code sandbox. And they powered the LM arena web arena. So it's basically the, just like you do LMS, like text to text, they do the same for like UI generation. So if you're asking a model, how do you do it? But yeah, I think that's kind of where.Dharmesh [00:48:45]: That's the thing I'm really fascinated by. So the early LLM, you know, we're understandably, but laughably bad at simple arithmetic, right? That's the thing like my wife, Normies would ask us, like, you call this AI, like it can't, my son would be like, it's just stupid. It can't even do like simple arithmetic. And then like we've discovered over time that, and there's a reason for this, right? It's like, it's a large, there's, you know, the word language is in there for a reason in terms of what it's been trained on. It's not meant to do math, but now it's like, okay, well, the fact that it has access to a Python interpreter that I can actually call at runtime, that solves an entire body of problems that it wasn't trained to do. And it's basically a form of delegation. And so the thought that's kind of rattling around in my head is that that's great. So it's, it's like took the arithmetic problem and took it first. Now, like anything that's solvable through a relatively concrete Python program, it's able to do a bunch of things that I couldn't do before. Can we get to the same place with UI? I don't know what the future of UI looks like in a agentic AI world, but maybe let the LLM handle it, but not in the classic sense. Maybe it generates it on the fly, or maybe we go through some iterations and hit cache or something like that. So it's a little bit more predictable. Uh, I don't know, but yeah.Alessio [00:49:48]: And especially when is the human supposed to intervene? So, especially if you're composing them, most of them should not have a UI because then they're just web hooking to somewhere else. I just want to touch back. I don't know if you have more comments on this.swyx [00:50:01]: I was just going to ask when you, you said you got, you're going to go back to code. What

Eye On A.I.
#244 Yoav Shoham on Jamba Models, Maestro and The Future of Enterprise AI

Eye On A.I.

Play Episode Listen Later Mar 27, 2025 52:16


This episode is sponsored by the DFINITY Foundation.  DFINITY Foundation's mission is to develop and contribute technology that enables the Internet Computer (ICP) blockchain and its ecosystem, aiming to shift cloud computing into a fully decentralized state.   Find out more at https://internetcomputer.org/ In this episode of Eye on AI, Yoav Shoham, co-founder of AI21 Labs, shares his insights on the evolution of AI, touching on key advancements such as Jamba and Maestro. From the early days of his career to the latest developments in AI systems, Yoav offers a comprehensive look into the future of artificial intelligence.   Yoav opens up about his journey in AI, beginning with his academic roots in game theory and logic, followed by his entrepreneurial ventures that led to the creation of AI21 Labs. He explains the founding of AI21 Labs and the company's mission to combine traditional AI approaches with modern deep learning methods, leading to innovations like Jamba—a highly efficient hybrid AI model that's disrupting the traditional transformer architecture.   He also introduces Maestro, AI21's orchestrator that works with multiple large language models (LLMs) and AI tools to create more reliable, predictable, and efficient systems for enterprises. Yoav discusses how Maestro is tackling real-world challenges in enterprise AI, moving beyond flashy demos to practical, scalable solutions.   Throughout the conversation, Yoav emphasizes the limitations of current large language models (LLMs), even those with reasoning capabilities, and explains how AI systems, rather than just pure language models, are becoming the future of AI. He also delves into the philosophical side of AI, discussing whether models truly "understand" and what that means for the future of artificial intelligence.   Whether you're deeply invested in AI research or curious about its applications in business, this episode is filled with valuable insights into the current and future landscape of artificial intelligence.     Stay Updated: Craig Smith Twitter: https://twitter.com/craigss Eye on A.I. Twitter: https://twitter.com/EyeOn_AI     (00:00) Introduction: The Future of AI Systems (02:33) Yoav's Journey: From Academia to AI21 Labs (05:57) The Evolution of AI: Symbolic AI and Deep Learning (07:38) Jurassic One: AI21 Labs' First Language Model (10:39) Jamba: Revolutionizing AI Model Architecture (16:11) Benchmarking AI Models: Challenges and Criticisms (22:18) Reinforcement Learning in AI Models (24:33) The Future of AI: Is Jamba the End of Larger Models? (27:31) Applications of Jamba: Real-World Use Cases in Enterprise (29:56) The Transition to Mass AI Deployment in Enterprises (33:47) Maestro: The Orchestrator of AI Tools and Language Models (36:03) GPT-4.5 and Reasoning Models: Are They the Future of AI? (38:09) Yoav's Pet Project: The Philosophical Side of AI Understanding (41:27) The Philosophy of AI Understanding (45:32) Explanations and Competence in AI (48:59) Where to Access Jamba and Maestro  

Machine Learning Street Talk
GSMSymbolic paper - Iman Mirzadeh (Apple)

Machine Learning Street Talk

Play Episode Listen Later Mar 19, 2025 71:23


Iman Mirzadeh from Apple, who recently published the GSM-Symbolic paper discusses the crucial distinction between intelligence and achievement in AI systems. He critiques current AI research methodologies, highlighting the limitations of Large Language Models (LLMs) in reasoning and knowledge representation. SPONSOR MESSAGES:***Tufa AI Labs is a brand new research lab in Zurich started by Benjamin Crouzier focussed on o-series style reasoning and AGI. They are hiring a Chief Engineer and ML engineers. Events in Zurich. Goto https://tufalabs.ai/***TRANSCRIPT + RESEARCH:https://www.dropbox.com/scl/fi/mlcjl9cd5p1kem4l0vqd3/IMAN.pdf?rlkey=dqfqb74zr81a5gqr8r6c8isg3&dl=0TOC:1. Intelligence vs Achievement in AI Systems [00:00:00] 1.1 Intelligence vs Achievement Metrics in AI Systems [00:03:27] 1.2 AlphaZero and Abstract Understanding in Chess [00:10:10] 1.3 Language Models and Distribution Learning Limitations [00:14:47] 1.4 Research Methodology and Theoretical Frameworks2. Intelligence Measurement and Learning [00:24:24] 2.1 LLM Capabilities: Interpolation vs True Reasoning [00:29:00] 2.2 Intelligence Definition and Measurement Approaches [00:34:35] 2.3 Learning Capabilities and Agency in AI Systems [00:39:26] 2.4 Abstract Reasoning and Symbol Understanding3. LLM Performance and Evaluation [00:47:15] 3.1 Scaling Laws and Fundamental Limitations [00:54:33] 3.2 Connectionism vs Symbolism Debate in Neural Networks [00:58:09] 3.3 GSM-Symbolic: Testing Mathematical Reasoning in LLMs [01:08:38] 3.4 Benchmark Evaluation and Model Performance AssessmentREFS:[00:01:00] AlphaZero chess AI system, Silver et al.https://arxiv.org/abs/1712.01815[00:07:10] Game Changer: AlphaZero's Groundbreaking Chess Strategies, Sadler & Reganhttps://www.amazon.com/Game-Changer-AlphaZeros-Groundbreaking-Strategies/dp/9056918184[00:11:35] Cross-entropy loss in language modeling, Voitahttp://lena-voita.github.io/nlp_course/language_modeling.html[00:17:20] GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in LLMs, Mirzadeh et al.https://arxiv.org/abs/2410.05229[00:21:25] Connectionism and Cognitive Architecture: A Critical Analysis, Fodor & Pylyshynhttps://www.sciencedirect.com/science/article/pii/001002779090014B[00:28:55] Brain-to-body mass ratio scaling laws, Sutskeverhttps://www.theverge.com/2024/12/13/24320811/what-ilya-sutskever-sees-openai-model-data-training[00:29:40] On the Measure of Intelligence, Chollethttps://arxiv.org/abs/1911.01547[00:33:30] On definition of intelligence, Gignac et al.https://www.sciencedirect.com/science/article/pii/S0160289624000266[00:35:30] Defining intelligence, Wanghttps://cis.temple.edu/~wangp/papers.html[00:37:40] How We Learn: Why Brains Learn Better Than Any Machine... for Now, Dehaenehttps://www.amazon.com/How-We-Learn-Brains-Machine/dp/0525559884[00:39:35] Surfaces and Essences: Analogy as the Fuel and Fire of Thinking, Hofstadter and Sanderhttps://www.amazon.com/Surfaces-Essences-Analogy-Fuel-Thinking/dp/0465018475[00:43:15] Chain-of-thought prompting, Wei et al.https://arxiv.org/abs/2201.11903[00:47:20] Test-time scaling laws in machine learning, Brownhttps://podcasts.apple.com/mv/podcast/openais-noam-brown-ilge-akkaya-and-hunter-lightman-on/id1750736528?i=1000671532058[00:47:50] Scaling Laws for Neural Language Models, Kaplan et al.https://arxiv.org/abs/2001.08361[00:55:15] Tensor product variable binding, Smolenskyhttps://www.sciencedirect.com/science/article/abs/pii/000437029090007M[01:08:45] GSM-8K dataset, OpenAIhttps://huggingface.co/datasets/openai/gsm8k

Pretty Curious with Jonathan Van Ness
You Might Also Like: On Purpose with Jay Shetty

Pretty Curious with Jonathan Van Ness

Play Episode Listen Later Mar 17, 2025


Introducing Deepak Chopra: How to Ask AI The RIGHT Questions To Grow, Heal, and Live a More Fulfilled Life from On Purpose with Jay Shetty.Follow the show: On Purpose with Jay ShettyDo you ever feel like you're just going through the motions? Do you ever feel lonely even when you're around people? Today, Jay welcomes back the legendary Deepak Chopra after six years, to discuss the unexpected intersection of spirituality and artificial intelligence. Together they unpack the beautiful blend of ancient wisdom and cutting-edge innovation. Jay reflects on his first meeting with Deepak and how their bond has evolved over the years. The conversation quickly flows into one of the most fascinating questions of our time: Can AI actually support our spiritual growth, rather than distract from it? Deepak shares his insights on the mysteries of the universe, the unseen forces that shape our reality, and why he believes consciousness, not matter, is the foundation of everything. Jay and Deepak break down the fears surrounding AI, from misinformation to war, and why it’s so important for us to grow spiritually alongside our technology. They explore how storytelling, creativity, and love are deeply human qualities that no machine can replicate, and how we can use these gifts to shape a better, more compassionate world. In this interview, you'll learn: How to Ask AI Better Questions to Unlock Wisdom How to Use Technology to Support Emotional and Physical Healing How to Find Purpose Through the Art of Storytelling How to Train AI to Reflect Inclusivity and Diversity How to Embrace Consciousness as the Foundation of Reality How to Create a Sacred Relationship with Technology Deepak reminds us that the answers we’re looking for often begin with a better question. And with the right mindset, even the most advanced technology can become a tool for inner transformation and collective healing. With Love and Gratitude, Jay Shetty Join over 750,000 people to receive my most transformative wisdom directly in your inbox every single week with my free newsletter. Subscribe here. Join Jay for his first ever, On Purpose Live Tour! Tickets are on sale now. Hope to see you there! What We Discuss: 00:00 Intro 01:41 What If the Universe Is Just a Giant Digital Simulation? 12:36 How to Train AI to Unlock Ancient and Hidden Knowledge 13:39 Blending AI and Spirituality to Understand Consciousness 18:23 Could AI Really Lead to Human Extinction? 22:34 What’s Actually Holding Humanity Back From Progress? 23:37 How the Human Brain Transformed Over Time 26:11 The 2 Things That Set Humans Apart From All Other Species 27:16 Can Technology Lead Us to True Peace and Prosperity? 28:10 Will AI Replace Our Jobs or Unlock Human Creativity? 30:46 Do You Think AI Can Ever Have a Soul? 31:45 The Gender and Racial Bias Hidden in AI Systems 32:33 How to Build More Inclusive and Equitable AI Models 33:32 Why a Shared Vision Can Solve Any Problem We Face 36:13 Would You Trust AI to Know You Personally? 36:57 How You can Use AI to Get Better Sleep 37:29 Can AI Actually Give You Good Relationship Advice? 38:07 How AI Can Help You Find and Nurture Love 38:29 Why Personal Growth Solutions Should Never Be Generic 39:41 Your DNA Holds the Footprints of Human History 42:33 Rethinking the Big Bang: What Science Still Can’t Explain 44:31 Is Everything You See Just a Projection? 47:48 Why Fear of the Unknown Limits Our Growth 48:52 Want Better Answers? Ask Better Questions 51:41 The True Secret to Longevity Isn’t What You Think 54:30 How Your Brain Turns Experience Into Reality 55:08 Why Consciousness Is Still Life’s Greatest Mystery 56:28 The First Question You Should Always Ask AI 58:38 How ChatGPT Can Spark Deeper, More Intelligent Questions Episode Resources: Deepak Chopra | Website Deepak Chopra | Instagram Deepak Chopra | TikTok Deepak Chopra | Facebook Deepak Chopra | YouTube Deepak Chopra | X Digital Dharma: How AI Can Elevate Spiritual Intelligence and Personal Well-BeingSee omnystudio.com/listener for privacy information.DISCLAIMER: Please note, this is an independent podcast episode not affiliated with, endorsed by, or produced in conjunction with the host podcast feed or any of its media entities. The views and opinions expressed in this episode are solely those of the creators and guests. For any concerns, please reach out to team@podroll.fm.

The Bootstrapped Founder
381: How AI Changes Famous Laws in Software and Entrepreneurship

The Bootstrapped Founder

Play Episode Listen Later Mar 14, 2025 24:59 Transcription Available


The rise of AI is fundamentally changing and challenging the classic laws and principles of software development and entrepreneurship. Drawing from my experience building Podscan.fm with AI assistance, I dive into how laws like Conway's Law, Brooks' Law, and Postel's Law are being transformed in this new era of AI-assisted development, while sharing practical insights for founders and developers navigating this shifting landscape.The blog post: https://thebootstrappedfounder.com/how-ai-changes-famous-laws-in-software-and-entrepreneurship/The podcast episode: https://tbf.fm/episodes/381-how-ai-changes-famous-laws-in-software-and-entrepreneurshipCheck out Podscan to get alerts when you're mentioned on podcasts: https://podscan.fmSend me a voicemail on Podline: https://podline.fm/arvidYou'll find my weekly article on my blog: https://thebootstrappedfounder.comPodcast: https://thebootstrappedfounder.com/podcastNewsletter: https://thebootstrappedfounder.com/newsletterMy book Zero to Sold: https://zerotosold.com/My book The Embedded Entrepreneur: https://embeddedentrepreneur.com/My course Find Your Following: https://findyourfollowing.comHere are a few tools I use. Using my affiliate links will support my work at no additional cost to you.- Notion (which I use to organize, write, coordinate, and archive my podcast + newsletter): https://affiliate.notion.so/465mv1536drx- Riverside.fm (that's what I recorded this episode with): https://riverside.fm/?via=arvid- TweetHunter (for speedy scheduling and writing Tweets): http://tweethunter.io/?via=arvid- HypeFury (for massive Twitter analytics and scheduling): https://hypefury.com/?via=arvid60- AudioPen (for taking voice notes and getting amazing summaries): https://audiopen.ai/?aff=PXErZ- Descript (for word-based video editing, subtitles, and clips): https://www.descript.com/?lmref=3cf39Q- ConvertKit (for email lists, newsletters, even finding sponsors): https://convertkit.com?lmref=bN9CZw

Let's Talk AI
#202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors

Let's Talk AI

Play Episode Listen Later Mar 9, 2025 79:52 Transcription Available


Our 202nd episode with a summary and discussion of last week's big AI news! Recorded on 03/07/2025 Hosted by Andrey Kurenkov and Jeremie Harris. Feel free to email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai Read out our text newsletter and comment on the podcast at https://lastweekin.ai/. Join our Discord here! https://discord.gg/nTyezGSKwP In this episode: Alibaba released Qwen-32B, their latest reasoning model, on par with leading models like DeepMind's R1. Anthropic raised $3.5 billion in a funding round, valuing the company at $61.5 billion, solidifying its position as a key competitor to OpenAI. DeepMind introduced BigBench Extra Hard, a more challenging benchmark to evaluate the reasoning capabilities of large language models. Reinforcement Learning pioneers Andrew Bartow and Rich Sutton were awarded the prestigious Turing Award for their contributions to the field. Timestamps + Links: cle picks: (00:00:00) Intro / Banter (00:01:41) Episode Preview (00:02:50) GPT-4.5 Discussion (00:14:13) Alibaba's New QwQ 32B Model is as Good as DeepSeek-R1 ; Outperforms OpenAI's o1-mini (00:21:29) With Alexa Plus, Amazon finally reinvents its best product (00:26:08) Another DeepSeek moment? General AI agent Manus shows ability to handle complex tasks (00:29:14) Microsoft's new Dragon Copilot is an AI assistant for healthcare (00:32:24) Mistral's new OCR API turns any PDF document into an AI-ready Markdown file (00:33:19) A.I. Start-Up Anthropic Closes Deal That Values It at $61.5 Billion (00:35:49) Nvidia-Backed CoreWeave Files for IPO, Shows Growing Revenue (00:38:05) Waymo and Uber's Austin robotaxi expansion begins today (00:38:54) UK competition watchdog drops Microsoft-OpenAI probe (00:41:17) Scale AI announces multimillion-dollar defense deal, a major step in U.S. military automation (00:44:43) DeepSeek Open Source Week: A Complete Summary (00:45:25) DeepSeek AI Releases DualPipe: A Bidirectional Pipeline Parallelism Algorithm for Computation-Communication Overlap in V3/R1 Training (00:53:00) Physical Intelligence open-sources Pi0 robotics foundation model (00:54:23) BIG-Bench Extra Hard (00:56:10) Cognitive Behaviors that Enable Self-Improving Reasoners (01:01:49) The MASK Benchmark: Disentangling Honesty From Accuracy in AI Systems (01:05:32) Pioneers of Reinforcement Learning Win the Turing Award (01:06:56) OpenAI launches $50M grant program to help fund academic research (01:07:25) The Nuclear-Level Risk of Superintelligent AI (01:13:34) METR's GPT-4.5 pre-deployment evaluations (01:17:16) Chinese buyers are getting Nvidia Blackwell chips despite US export controls

Times Higher Education
Campus podcast: Why we need interdisciplinarity in teaching and research

Times Higher Education

Play Episode Listen Later Mar 6, 2025 56:52


Complex problems cannot be solved if examined only through a narrow lens. Enter interdisciplinarity. It is now widely accepted that drawing on varied expertise and perspectives is the only way we can understand and tackle many of the most challenging issues we face, as individuals and as a species. So, there is a growing movement towards more cross disciplinary working in higher education but it faces challenges. Interdisciplinarity requires a shift of mindset in an academy built upon clear disciplinary distinctions and must compete for space in already overcrowded curricula. We speak to two leadings scholars in interdisciplinary research and teaching to find out why it is so important and how they are encouraging more academics and students to break out of traditional academic silos. Gabriele Bammer is a professor of integration and implementation sciences (i2S) at the Australian National University. She is author of several books including ‘Disciplining Interdisciplinarity' and is inaugural president of the Global Alliance for Inter- and Transdisciplinarity. To support progress in interdisciplinarity around the world, she runs the Integration and Implementation Insights blog and repository of theory, methods and tools underpinning i2S. Gabriele has held visiting appointments at Harvard University's John F. Kennedy School of Government, the National Socio-Environmental Synthesis Center at the University of Maryland and the Institute for Advanced Sustainability Studies in Potsdam, Germany. Kate Crawford is an international scholar of the social implications of artificial intelligence who has advised policymakers in the United Nations, the White House, and the European Parliament on AI, and currently leads the Knowing Machines Project, an international research collaboration that investigates the foundations of machine learning. She is a research professor at USC Annenberg in Los Angeles, a senior principal researcher at MSR in New York, an honorary professor at the University of Sydney, and the inaugural visiting chair for AI and Justice at the École Normale Supérieure in Paris. Her award-winning book, Atlas of AI, reveals the extractive nature of this technology while her creative collaborations such as Anatomy of an AI System with Vladan Joler and Excavating AI with Trevor Paglen explore the complex processes behind each human-AI interaction, showing the material and human costs. Her latest exhibition, Calculating Empires: A Genealogy of Technology and Power 1500-2025, opened in Milan, November 2023 and won the Grand Prize of the European Commission for art and technology. More advice and insight can be found in our latest Campus spotlight guide: A focus on interdisciplinarity in teaching.

SemiWiki.com
Podcast EP276: How Alphawave Semi is Fueling the Next Generation of AI Systems with Letizia Giuliano

SemiWiki.com

Play Episode Listen Later Feb 28, 2025 16:06


Dan is joined by Letizia Giuliano, Vice President of Product Marketing and Management at Alphawave Semi. She specializes in architecting cutting-edge solutions for high-speed connectivity and chiplet design architecture. Prior to her role at Alphawave Semi, Letizia held the position of Product Line Manager at Intel, where… Read More

Shift AI Podcast
Reshaping Workplace Communication with AI, Textio Co-Founder Kieran Snyder

Shift AI Podcast

Play Episode Listen Later Feb 20, 2025 36:52


In this episode of the Shift AI Podcast, Kieran Snyder, co-founder of Textio, joins Boaz Ashkenazy to explore the evolving landscape of AI in workplace communications. From her early days combining linguistics and engineering to pioneering bias detection in HR technology, Snyder shares unique insights into the responsible development of AI technologies. The conversation delves into the challenges of bias in large language models, the future of enterprise AI adoption, and observations about how different generations approach AI tools. Drawing from her experience early pre-ChatGPT development and her recent survey of students' AI usage, Snyder offers a thoughtful perspective on the intersection of AI and human labor. Whether you're interested in the future of workplace communication, the challenges of bias in AI, or how the next generation will reshape our relationship with technology, this episode provides valuable insights from someone who has been at the forefront of AI innovation for over a decade.Chapters:[00:00] The Current State of AI Adoption[02:04] Origins and Background: From Engineering to Entrepreneurship[06:52] The Evolution of Textio: From Word Processing to HR Tech[09:36] Generative AI and Early Industry Insights[14:26] Addressing Bias in AI Systems[17:41] The AI Rush: Challenges in Enterprise Adoption[22:12] The Next Generation's Perspective on AI[26:17] AI Policy and Regulation[28:50] Mentorship and the Future of Work[34:33] Looking Ahead: The Actively Emerging FutureConnect with Kieran SnyderLinkedIn: https://www.linkedin.com/in/kieran-snyder/ Connect with Boaz AshkenazyLinkedIn: https://www.linkedin.com/in/boazashkenazy X: boazashkenazyEmail: shift@augmentedailabs.com

Last Call with Steve Noviello: The Podcast
S2E83 Would you trust an AI system to self-drive your car?

Last Call with Steve Noviello: The Podcast

Play Episode Listen Later Feb 17, 2025 22:15


IT'S LAST CALL, FIRST SHOT WEEK.Steve Noviello is joined by Steve Eager, Paige Ellenberger and Brian Demaris to discuss envelope etiquette, a cry for help on social media, and the debate of appropriate tipping or over tipping.

Machine Learning Street Talk
Sepp Hochreiter - LSTM: The Comeback Story?

Machine Learning Street Talk

Play Episode Listen Later Feb 12, 2025 67:01


Sepp Hochreiter, the inventor of LSTM (Long Short-Term Memory) networks – a foundational technology in AI. Sepp discusses his journey, the origins of LSTM, and why he believes his latest work, XLSTM, could be the next big thing in AI, particularly for applications like robotics and industrial simulation. He also shares his controversial perspective on Large Language Models (LLMs) and why reasoning is a critical missing piece in current AI systems.SPONSOR MESSAGES:***CentML offers competitive pricing for GenAI model deployment, with flexible options to suit a wide range of models, from small to large-scale deployments. Check out their super fast DeepSeek R1 hosting!https://centml.ai/pricing/Tufa AI Labs is a brand new research lab in Zurich started by Benjamin Crouzier focussed on o-series style reasoning and AGI. They are hiring a Chief Engineer and ML engineers. Events in Zurich.Goto https://tufalabs.ai/***TRANSCRIPT AND BACKGROUND READING:https://www.dropbox.com/scl/fi/n1vzm79t3uuss8xyinxzo/SEPPH.pdf?rlkey=fp7gwaopjk17uyvgjxekxrh5v&dl=0Prof. Sepp Hochreiterhttps://www.nx-ai.com/https://x.com/hochreitersepphttps://scholar.google.at/citations?user=tvUH3WMAAAAJ&hl=enTOC:1. LLM Evolution and Reasoning Capabilities[00:00:00] 1.1 LLM Capabilities and Limitations Debate[00:03:16] 1.2 Program Generation and Reasoning in AI Systems[00:06:30] 1.3 Human vs AI Reasoning Comparison[00:09:59] 1.4 New Research Initiatives and Hybrid Approaches2. LSTM Technical Architecture[00:13:18] 2.1 LSTM Development History and Technical Background[00:20:38] 2.2 LSTM vs RNN Architecture and Computational Complexity[00:25:10] 2.3 xLSTM Architecture and Flash Attention Comparison[00:30:51] 2.4 Evolution of Gating Mechanisms from Sigmoid to Exponential3. Industrial Applications and Neuro-Symbolic AI[00:40:35] 3.1 Industrial Applications and Fixed Memory Advantages[00:42:31] 3.2 Neuro-Symbolic Integration and Pi AI Project[00:46:00] 3.3 Integration of Symbolic and Neural AI Approaches[00:51:29] 3.4 Evolution of AI Paradigms and System Thinking[00:54:55] 3.5 AI Reasoning and Human Intelligence Comparison[00:58:12] 3.6 NXAI Company and Industrial AI ApplicationsREFS:[00:00:15] Seminal LSTM paper establishing Hochreiter's expertise (Hochreiter & Schmidhuber)https://direct.mit.edu/neco/article-abstract/9/8/1735/6109/Long-Short-Term-Memory[00:04:20] Kolmogorov complexity and program composition limitations (Kolmogorov)https://link.springer.com/article/10.1007/BF02478259[00:07:10] Limitations of LLM mathematical reasoning and symbolic integration (Various Authors)https://www.arxiv.org/pdf/2502.03671[00:09:05] AlphaGo's Move 37 demonstrating creative AI (Google DeepMind)https://deepmind.google/research/breakthroughs/alphago/[00:10:15] New AI research lab in Zurich for fundamental LLM research (Benjamin Crouzier)https://tufalabs.ai[00:19:40] Introduction of xLSTM with exponential gating (Beck, Hochreiter, et al.)https://arxiv.org/abs/2405.04517[00:22:55] FlashAttention: fast & memory-efficient attention (Tri Dao et al.)https://arxiv.org/abs/2205.14135[00:31:00] Historical use of sigmoid/tanh activation in 1990s (James A. McCaffrey)https://visualstudiomagazine.com/articles/2015/06/01/alternative-activation-functions.aspx[00:36:10] Mamba 2 state space model architecture (Albert Gu et al.)https://arxiv.org/abs/2312.00752[00:46:00] Austria's Pi AI project integrating symbolic & neural AI (Hochreiter et al.)https://www.jku.at/en/institute-of-machine-learning/research/projects/[00:48:10] Neuro-symbolic integration challenges in language models (Diego Calanzone et al.)https://openreview.net/forum?id=7PGluppo4k[00:49:30] JKU Linz's historical and neuro-symbolic research (Sepp Hochreiter)https://www.jku.at/en/news-events/news/detail/news/bilaterale-ki-projekt-unter-leitung-der-jku-erhaelt-fwf-cluster-of-excellence/YT: https://www.youtube.com/watch?v=8u2pW2zZLCs

Discover Daily by Perplexity
Altman Reconsiders Open-Source Strategy, EU Bans Risky AI Systems, and Super-Earth Discovered

Discover Daily by Perplexity

Play Episode Listen Later Feb 7, 2025 9:09 Transcription Available


We're experimenting and would love to hear from you!In this episode of 'Discover Daily', we explore Sam Altman's recent acknowledgment that OpenAI may need to reconsider its open-source strategy, suggesting they might be "on the wrong side of history." This significant shift comes as competitive pressure from open-source models like DeepSeek R1 continues to mount, with Altman praising DeepSeek as "a very good model" that has narrowed OpenAI's traditional lead in the field.The European Union has taken a historic step in AI regulation with the implementation of the EU AI Act's first phase on February 2, 2025. The legislation prohibits AI systems deemed to pose "unacceptable risks," including manipulative systems, social scoring, and untargeted facial recognition databases. Violations can result in substantial penalties of up to €35 million or 7% of a company's total worldwide annual turnover, demonstrating the EU's commitment to establishing itself as a global leader in trustworthy AI development.Our main story focuses on two remarkable super-Earth discoveries within their stars' habitable zones. TOI-715 b, located 137 light-years away, is approximately 1.5 times wider than Earth and orbits its red dwarf star every 19 days. The second discovery, HD 20794 d, orbits a Sun-like star just 20 light-years from Earth and is roughly six times more massive than Earth, with an elliptical orbit that moves in and out of the habitable zone. These discoveries represent significant milestones in our search for potentially habitable worlds and provide promising targets for future research with advanced instruments like the James Webb Space Telescope.From Perplexity's Discover Feed: https://www.perplexity.ai/page/altman-reconsiders-open-source-fT0uV12jTna0XkxW8xkEDQ https://www.perplexity.ai/page/eu-bans-risky-ai-systems-.iTygUNvS2mKll.lL9xFdAhttps://www.perplexity.ai/page/super-earth-discovered-WR42RfwCSQWU1ebaQaeQxw Perplexity is the fastest and most powerful way to search the web. Perplexity crawls the web and curates the most relevant and up-to-date sources (from academic papers to Reddit threads) to create the perfect response to any question or topic you're interested in. Take the world's knowledge with you anywhere. Available on iOS and Android Join our growing Discord community for the latest updates and exclusive content. Follow us on: Instagram Threads X (Twitter) YouTube Linkedin

The Legacy Leaders Show With Izabela Lundberg
From Air Force Officer to AI & Systems Engineering Innovator

The Legacy Leaders Show With Izabela Lundberg

Play Episode Listen Later Feb 6, 2025 62:54


Tune into the Legacy Leaders Show for an enlightening episode featuring Rachel Dzieran, a trailblazer in AI and healthcare and the founder and CEO of the Navy SEALs Fund. With a distinguished educational journey from the United States Air Force Academy to a PhD in Systems Engineering, Rachel leads the charge in pioneering medical technologies. This episode explores her revolutionary AI-driven methods for kidney transplant decision-making, shares powerful leadership lessons from her military and biotech experiences, and unveils her vision for the future of healthcare technology. By listening, you'll gain insights into the practical applications of AI in medicine, learn how disciplined leadership can foster technological innovation, and consider the ethical dimensions of integrating AI into healthcare practices. Additionally, we'll delve into her role with the Navy SEALs Fund, emphasizing her commitment to supporting veterans and their families. Join us for a conversation that promises to inspire and provide actionable knowledge for navigating the complexities of modern medical advancements.

Cyber Security Headlines
Meta identifies risky AI systems, Ferret malware joins 'Contagious Interview' campaign, credential theft rises as a target

Cyber Security Headlines

Play Episode Listen Later Feb 5, 2025 7:10


Meta says it may stop development of AI systems it deems too risky Ferret Malware Added to 'Contagious Interview' Campaign Credential Theft Becomes Cybercriminals' Favorite Target Huge thanks to our episode sponsor, ThreatLocker ThreatLocker® is a global leader in Zero Trust endpoint security, offering cybersecurity controls to protect businesses from zero-day attacks and ransomware. ThreatLocker operates with a default deny approach to reduce the attack surface and mitigate potential cyber vulnerabilities. To learn more and start your free trial, visit ThreatLocker.com. Find the stories behind the headlines at CISOseries.com.

80,000 Hours Podcast with Rob Wiblin
#132 Classic episode – Nova DasSarma on why information security may be critical to the safe development of AI systems

80,000 Hours Podcast with Rob Wiblin

Play Episode Listen Later Jan 31, 2025 161:11


If a business has spent $100 million developing a product, it's a fair bet that they don't want it stolen in two seconds and uploaded to the web where anyone can use it for free.This problem exists in extreme form for AI companies. These days, the electricity and equipment required to train cutting-edge machine learning models that generate uncanny human text and images can cost tens or hundreds of millions of dollars. But once trained, such models may be only a few gigabytes in size and run just fine on ordinary laptops.Today's guest, the computer scientist and polymath Nova DasSarma, works on computer and information security for the AI company Anthropic with the security team. One of her jobs is to stop hackers exfiltrating Anthropic's incredibly expensive intellectual property, as recently happened to Nvidia. Rebroadcast: this episode was originally released in June 2022.Links to learn more, highlights, and full transcript.As she explains, given models' small size, the need to store such models on internet-connected servers, and the poor state of computer security in general, this is a serious challenge.The worries aren't purely commercial though. This problem looms especially large for the growing number of people who expect that in coming decades we'll develop so-called artificial ‘general' intelligence systems that can learn and apply a wide range of skills all at once, and thereby have a transformative effect on society.If aligned with the goals of their owners, such general AI models could operate like a team of super-skilled assistants, going out and doing whatever wonderful (or malicious) things are asked of them. This might represent a huge leap forward for humanity, though the transition to a very different new economy and power structure would have to be handled delicately.If unaligned with the goals of their owners or humanity as a whole, such broadly capable models would naturally ‘go rogue,' breaking their way into additional computer systems to grab more computing power — all the better to pursue their goals and make sure they can't be shut off.As Nova explains, in either case, we don't want such models disseminated all over the world before we've confirmed they are deeply safe and law-abiding, and have figured out how to integrate them peacefully into society. In the first scenario, premature mass deployment would be risky and destabilising. In the second scenario, it could be catastrophic — perhaps even leading to human extinction if such general AI systems turn out to be able to self-improve rapidly rather than slowly, something we can only speculate on at this point.If highly capable general AI systems are coming in the next 10 or 20 years, Nova may be flying below the radar with one of the most important jobs in the world.We'll soon need the ability to ‘sandbox' (i.e. contain) models with a wide range of superhuman capabilities, including the ability to learn new skills, for a period of careful testing and limited deployment — preventing the model from breaking out, and criminals from breaking in. Nova and her colleagues are trying to figure out how to do this, but as this episode reveals, even the state of the art is nowhere near good enough.Chapters:Cold open (00:00:00)Rob's intro (00:00:52)The interview begins (00:02:44)Why computer security matters for AI safety (00:07:39)State of the art in information security (00:17:21)The hack of Nvidia (00:26:50)The most secure systems that exist (00:36:27)Formal verification (00:48:03)How organisations can protect against hacks (00:54:18)Is ML making security better or worse? (00:58:11)Motivated 14-year-old hackers (01:01:08)Disincentivising actors from attacking in the first place (01:05:48)Hofvarpnir Studios (01:12:40)Capabilities vs safety (01:19:47)Interesting design choices with big ML models (01:28:44)Nova's work and how she got into it (01:45:21)Anthropic and career advice (02:05:52)$600M Ethereum hack (02:18:37)Personal computer security advice (02:23:06)LastPass (02:31:04)Stuxnet (02:38:07)Rob's outro (02:40:18)Producer: Keiran HarrisAudio mastering: Ben Cordell and Beppe RådvikTranscriptions: Katy Moore

The Brave Marketer
Future-Proofing AI: Ethics, Accountability and Governance

The Brave Marketer

Play Episode Listen Later Jan 29, 2025 29:32


Jo Davaris, Global Chief Privacy Officer at Booking Holdings, discusses how to build trust in AI, emphasizing privacy and ethical governance as the pillars of AI innovation. We explore the importance of feedback mechanisms and transparency in AI systems, along with how to maintain user control without compromising innovation. Key Takeaways:  Building trust in AI systems and the role of privacy in AI governance The need for public and private collaboration on AI regulations Concerns with AI's rapid advancement The importance of integrating privacy and ethical considerations into AI development from the start Guest Bio:  Jo Davaris is the Global Chief Privacy Officer at Booking Holdings, where she established a unified privacy framework across brands like Booking.com and OpenTable. She leads AI Governance, Digital Trust, and Cybersecurity efforts, integrating privacy and innovation globally. Previously, Jo was Mercer's first Global Chief Privacy Officer, developing a privacy program that protected data while enabling growth. At American Express, she strengthened privacy governance for institutional & network businesses. Earlier, as an attorney for New York City's Administration for Children's Services, she prosecuted child abuse cases. Jo serves on advisory boards, speaks at conferences on privacy and governance, and advocates for social services in NYC. ---------------------------------------------------------------------------------------- About this Show: The Brave Technologist is here to shed light on the opportunities and challenges of emerging tech. To make it digestible, less scary, and more approachable for all! Join us as we embark on a mission to demystify artificial intelligence, challenge the status quo, and empower everyday people to embrace the digital revolution. Whether you're a tech enthusiast, a curious mind, or an industry professional, this podcast invites you to join the conversation and explore the future of AI together. The Brave Technologist Podcast is hosted by Luke Mulks, VP Business Operations at Brave Software—makers of the privacy-respecting Brave browser and Search engine, and now powering AI everywhere with the Brave Search API. Music by: Ari Dvorin Produced by: Sam Laliberte  

Eye On A.I.
#232 Sepp Hochreiter: How LSTMs Power Modern AI System's

Eye On A.I.

Play Episode Listen Later Jan 22, 2025 51:08


In this special episode of the Eye on AI podcast, Sepp Hochreiter, the inventor of Long Short-Term Memory (LSTM) networks, joins Craig Smith to discuss the profound impact of LSTMs on artificial intelligence, from language models to real-time robotics. Sepp reflects on the early days of LSTM development, sharing insights into his collaboration with Jürgen Schmidhuber and the challenges they faced in gaining recognition for their groundbreaking work. He explains how LSTMs became the foundation for technologies used by giants like Amazon, Apple, and Google, and how they paved the way for modern advancements like transformers. Topics include: - The origin story of LSTMs and their unique architecture. - Why LSTMs were crucial for sequence data like speech and text. - The rise of transformers and how they compare to LSTMs. - Real-time robotics: using LSTMs to build energy-efficient, autonomous systems. The next big challenges for AI and robotics in the era of generative AI. Sepp also shares his optimistic vision for the future of AI, emphasizing the importance of efficient, scalable models and their potential to revolutionize industries from healthcare to autonomous vehicles. Don't miss this deep dive into the history and future of AI, featuring one of its most influential pioneers. (00:00) Introduction: Meet Sepp Hochreiter (01:10) The Origins of LSTMs (02:26) Understanding the Vanishing Gradient Problem (05:12) Memory Cells and LSTM Architecture (06:35) Early Applications of LSTMs in Technology (09:38) How Transformers Differ from LSTMs (13:38) Exploring XLSTM for Industrial Applications (15:17) AI for Robotics and Real-Time Systems (18:55) Expanding LSTM Memory with Hopfield Networks (21:18) The Road to XLSTM Development (23:17) Industrial Use Cases of XLSTM (27:49) AI in Simulation: A New Frontier (32:26) The Future of LSTMs and Scalability (35:48) Inference Efficiency and Potential Applications (39:53) Continuous Learning and Adaptability in AI (42:59) Training Robots with XLSTM Technology (44:47) NXAI: Advancing AI in Industry

Machine Learning Street Talk
How Do AI Models Actually Think? - Laura Ruis

Machine Learning Street Talk

Play Episode Listen Later Jan 20, 2025 78:01


Laura Ruis, a PhD student at University College London and researcher at Cohere, explains her groundbreaking research into how large language models (LLMs) perform reasoning tasks, the fundamental mechanisms underlying LLM reasoning capabilities, and whether these models primarily rely on retrieval or develop procedural knowledge. SPONSOR MESSAGES: *** CentML offers competitive pricing for GenAI model deployment, with flexible options to suit a wide range of models, from small to large-scale deployments. https://centml.ai/pricing/ Tufa AI Labs is a brand new research lab in Zurich started by Benjamin Crouzier focussed on o-series style reasoning and AGI. Are you interested in working on reasoning, or getting involved in their events? Goto https://tufalabs.ai/ *** TOC 1. LLM Foundations and Learning 1.1 Scale and Learning in Language Models [00:00:00] 1.2 Procedural Knowledge vs Fact Retrieval [00:03:40] 1.3 Influence Functions and Model Analysis [00:07:40] 1.4 Role of Code in LLM Reasoning [00:11:10] 1.5 Semantic Understanding and Physical Grounding [00:19:30] 2. Reasoning Architectures and Measurement 2.1 Measuring Understanding and Reasoning in Language Models [00:23:10] 2.2 Formal vs Approximate Reasoning and Model Creativity [00:26:40] 2.3 Symbolic vs Subsymbolic Computation Debate [00:34:10] 2.4 Neural Network Architectures and Tensor Product Representations [00:40:50] 3. AI Agency and Risk Assessment 3.1 Agency and Goal-Directed Behavior in Language Models [00:45:10] 3.2 Defining and Measuring Agency in AI Systems [00:49:50] 3.3 Core Knowledge Systems and Agency Detection [00:54:40] 3.4 Language Models as Agent Models and Simulator Theory [01:03:20] 3.5 AI Safety and Societal Control Mechanisms [01:07:10] 3.6 Evolution of AI Capabilities and Emergent Risks [01:14:20] REFS: [00:01:10] Procedural Knowledge in Pretraining & LLM Reasoning Ruis et al., 2024 https://arxiv.org/abs/2411.12580 [00:03:50] EK-FAC Influence Functions in Large LMs Grosse et al., 2023 https://arxiv.org/abs/2308.03296 [00:13:05] Surfaces and Essences: Analogy as the Core of Cognition Hofstadter & Sander https://www.amazon.com/Surfaces-Essences-Analogy-Fuel-Thinking/dp/0465018475 [00:13:45] Wittgenstein on Language Games https://plato.stanford.edu/entries/wittgenstein/ [00:14:30] Montague Semantics for Natural Language https://plato.stanford.edu/entries/montague-semantics/ [00:19:35] The Chinese Room Argument David Cole https://plato.stanford.edu/entries/chinese-room/ [00:19:55] ARC: Abstraction and Reasoning Corpus François Chollet https://arxiv.org/abs/1911.01547 [00:24:20] Systematic Generalization in Neural Nets Lake & Baroni, 2023 https://www.nature.com/articles/s41586-023-06668-3 [00:27:40] Open-Endedness & Creativity in AI Tim Rocktäschel https://arxiv.org/html/2406.04268v1 [00:30:50] Fodor & Pylyshyn on Connectionism https://www.sciencedirect.com/science/article/abs/pii/0010027788900315 [00:31:30] Tensor Product Representations Smolensky, 1990 https://www.sciencedirect.com/science/article/abs/pii/000437029090007M [00:35:50] DreamCoder: Wake-Sleep Program Synthesis Kevin Ellis et al. https://courses.cs.washington.edu/courses/cse599j1/22sp/papers/dreamcoder.pdf [00:36:30] Compositional Generalization Benchmarks Ruis, Lake et al., 2022 https://arxiv.org/pdf/2202.10745 [00:40:30] RNNs & Tensor Products McCoy et al., 2018 https://arxiv.org/abs/1812.08718 [00:46:10] Formal Causal Definition of Agency Kenton et al. https://arxiv.org/pdf/2208.08345v2 [00:48:40] Agency in Language Models Sumers et al. https://arxiv.org/abs/2309.02427 [00:55:20] Heider & Simmel's Moving Shapes Experiment https://www.nature.com/articles/s41598-024-65532-0 [01:00:40] Language Models as Agent Models Jacob Andreas, 2022 https://arxiv.org/abs/2212.01681 [01:13:35] Pragmatic Understanding in LLMs Ruis et al. https://arxiv.org/abs/2210.14986

Machine Learning Street Talk
Yoshua Bengio - Designing out Agency for Safe AI

Machine Learning Street Talk

Play Episode Listen Later Jan 15, 2025 101:53


Professor Yoshua Bengio is a pioneer in deep learning and Turing Award winner. Bengio talks about AI safety, why goal-seeking “agentic” AIs might be dangerous, and his vision for building powerful AI tools without giving them agency. Topics include reward tampering risks, instrumental convergence, global AI governance, and how non-agent AIs could revolutionize science and medicine while reducing existential threats. Perfect for anyone curious about advanced AI risks and how to manage them responsibly. SPONSOR MESSAGES: *** CentML offers competitive pricing for GenAI model deployment, with flexible options to suit a wide range of models, from small to large-scale deployments. https://centml.ai/pricing/ Tufa AI Labs is a brand new research lab in Zurich started by Benjamin Crouzier focussed on o-series style reasoning and AGI. Are you interested in working on reasoning, or getting involved in their events? They are hosting an event in Zurich on January 9th with the ARChitects, join if you can. Goto https://tufalabs.ai/ *** Interviewer: Tim Scarfe Yoshua Bengio: https://x.com/Yoshua_Bengio https://scholar.google.com/citations?user=kukA0LcAAAAJ&hl=en https://yoshuabengio.org/ https://en.wikipedia.org/wiki/Yoshua_Bengio TOC: 1. AI Safety Fundamentals [00:00:00] 1.1 AI Safety Risks and International Cooperation [00:03:20] 1.2 Fundamental Principles vs Scaling in AI Development [00:11:25] 1.3 System 1/2 Thinking and AI Reasoning Capabilities [00:15:15] 1.4 Reward Tampering and AI Agency Risks [00:25:17] 1.5 Alignment Challenges and Instrumental Convergence 2. AI Architecture and Safety Design [00:33:10] 2.1 Instrumental Goals and AI Safety Fundamentals [00:35:02] 2.2 Separating Intelligence from Goals in AI Systems [00:40:40] 2.3 Non-Agent AI as Scientific Tools [00:44:25] 2.4 Oracle AI Systems and Mathematical Safety Frameworks 3. Global Governance and Security [00:49:50] 3.1 International AI Competition and Hardware Governance [00:51:58] 3.2 Military and Security Implications of AI Development [00:56:07] 3.3 Personal Evolution of AI Safety Perspectives [01:00:25] 3.4 AI Development Scaling and Global Governance Challenges [01:12:10] 3.5 AI Regulation and Corporate Oversight 4. Technical Innovations [01:23:00] 4.1 Evolution of Neural Architectures: From RNNs to Transformers [01:26:02] 4.2 GFlowNets and Symbolic Computation [01:30:47] 4.3 Neural Dynamics and Consciousness [01:34:38] 4.4 AI Creativity and Scientific Discovery SHOWNOTES (Transcript, references, best clips etc): https://www.dropbox.com/scl/fi/ajucigli8n90fbxv9h94x/BENGIO_SHOW.pdf?rlkey=38hi2m19sylnr8orb76b85wkw&dl=0 CORE REFS (full list in shownotes and pinned comment): [00:00:15] Bengio et al.: "AI Risk" Statement https://www.safe.ai/work/statement-on-ai-risk [00:23:10] Bengio on reward tampering & AI safety (Harvard Data Science Review) https://hdsr.mitpress.mit.edu/pub/w974bwb0 [00:40:45] Munk Debate on AI existential risk, featuring Bengio https://munkdebates.com/debates/artificial-intelligence [00:44:30] "Can a Bayesian Oracle Prevent Harm from an Agent?" (Bengio et al.) on oracle-to-agent safety https://arxiv.org/abs/2408.05284 [00:51:20] Bengio (2024) memo on hardware-based AI governance verification https://yoshuabengio.org/wp-content/uploads/2024/08/FlexHEG-Memo_August-2024.pdf [01:12:55] Bengio's involvement in EU AI Act code of practice https://digital-strategy.ec.europa.eu/en/news/meet-chairs-leading-development-first-general-purpose-ai-code-practice [01:27:05] Complexity-based compositionality theory (Elmoznino, Jiralerspong, Bengio, Lajoie) https://arxiv.org/abs/2410.14817 [01:29:00] GFlowNet Foundations (Bengio et al.) for probabilistic inference https://arxiv.org/pdf/2111.09266 [01:32:10] Discrete attractor states in neural systems (Nam, Elmoznino, Bengio, Lajoie) https://arxiv.org/pdf/2302.06403

Heads Talk
237 - Ludwig von Reiche, MD: Cyber Protect Series, NVIDIA - Sovereign AI Systems and Next-Level Parallel Computing

Heads Talk

Play Episode Listen Later Dec 16, 2024 32:00