POPULARITY
(0:00) Intro (1:49) About the podcast sponsor: The American College of Governance Counsel(2:36) Introduction by Professor Anat Admati, Stanford Graduate School of Business. Read the event coverage from Stanford's CASI.(4:14) Start of Interview(4:45) What inspired Karen to write this book and how she got started with journalism.(8:00) OpenAI's Nonprofit Origin Story(8:45) Sam Altman and Elon Musk's Collaboration(10:39) The Shift to For-Profit(12:12) On the original split between Musk and Altman over control of OpenAI(14:36) The Concept of AI Empires(18:04) About concept of "benefit to humanity" and OpenAI's mission "to ensure that AGI benefits all of humanity"(20:30) On Sam Altman's Ouster and OpenAI's Boardroom Drama (Nov 2023) "Doomers vs Boomers"(26:05) Investor Dynamics Post-Ouster of Sam Altman(28:21) Prominent Departures from OpenAI (ie Elon Musk, Dario Amodei, Ilya Sutskever, Mira Murati, etc)(30:55) The Geopolitics of AI: U.S. vs. China(32:37) The "What about China" Card used by US companies to ward off regulation.(34:26) "Scaling at All Costs is not leading us in a good place"(36:46) Karen's preference on ethical AI development "I really want there to be more participatory AI development. And I think about the full supply chain of AI development when I say that."(39:53) Her biggest hope and fear for the future "the greatest threat of these AI empires is the erosion of democracy."(43:34) The case of Chilean Community Activism and Empowerment(47:20) Recreating human intelligence and the example of Joseph Weizenbaum, MIT (Computer Power and Human Reason, 1976)(51:15) OpenAI's current AI research capabilities: "I think it's asymptotic because they have started tapping out of their scaling paradigm"(53:26) The state (and importance of) open source development of AI. "We need things to be more open"(55:08) The Bill Gates demo on chatGPT acing the AP Biology test.(58:54) Funding academic AI research and the public policy question on the role of Government.(1:01:11) Recommendations for Startups and UniversitiesKaren Hao is the author of Empire of AI (Penguin Press, May 2025) and an award-winning journalist covering the intersections of AI & society. You can follow Evan on social media at:X: @evanepsteinLinkedIn: https://www.linkedin.com/in/epsteinevan/ Substack: https://evanepstein.substack.com/__To support this podcast you can join as a subscriber of the Boardroom Governance Newsletter at https://evanepstein.substack.com/__Music/Soundtrack (found via Free Music Archive): Seeing The Future by Dexter Britain is licensed under a Attribution-Noncommercial-Share Alike 3.0 United States License
Audio note: this article contains 127 uses of latex notation, so the narration may be difficult to follow. There's a link to the original text in the episode description. Confidence: Medium, underlying data is patchy and relies on a good amount of guesswork, data work involved a fair amount of vibecoding. Intro: Tom Davidson has an excellent post explaining the compute bottleneck objection to the software-only intelligence explosion.[1] The rough idea is that AI research requires two inputs: cognitive labor and research compute. If these two inputs are gross complements, then even if there is recursive self-improvement in the amount of cognitive labor directed towards AI research, this process will fizzle as you get bottlenecked by the amount of research compute. The compute bottleneck objection to the software-only intelligence explosion crucially relies on compute and cognitive labor being gross complements; however, this fact is not [...] ---Outline:(00:35) Intro:(02:16) Model(02:19) Baseline CES in Compute(04:07) Conditions for a Software-Only Intelligence Explosion(07:39) Deriving the Estimation Equation(09:31) Alternative CES Formulation in Frontier Experiments(10:59) Estimation(11:02) Data(15:02) Trends(15:58) Estimation Results(18:52) ResultsThe original text contained 13 footnotes which were omitted from this narration. --- First published: June 1st, 2025 Source: https://forum.effectivealtruism.org/posts/xoX936hEvpxToeuLw/estimating-the-substitutability-between-compute-and --- Narrated by TYPE III AUDIO. ---Images from the article:Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
In this episode of Catching Up, Nate McClennen and Mason Pashia explore the evolving landscape of education with a focus on AI leadership, applied humanities, and innovative learning spaces. They dive into key topics such as the role of AI in transforming educational practices, the shift towards applied humanities in higher education, and the potential of community spaces in schools to enhance learning experiences. With insights from recent developments and thoughtful discussions, this episode offers valuable perspectives for educators, administrators, and policymakers looking to navigate the future of education. Tune in to discover how these emerging trends are shaping new pathways for learners and educators alike. Outline (00:00) Introduction and Episode Overview (02:40) Educational Innovations and Legislation (04:41) Future of Learning: Digital Wallets and AI (10:16) Applied Humanities in Higher Education (18:20) AI in Schools: Leadership, Crowd, and Lab (28:45) Reimagining Community Spaces in Schools (33:31) AI in Job Screening and Hiring (42:25) What's That Song? Links Watch the full video Read the full blog here Mason's LER Blog Ministry of Imagination Accreditation MSA Teach For America Should Embrace Apprenticeship Model Amid AmeriCorps Cuts LER digital wallet Michigan experiment Educational Choice for Children Act H.R 3250 Applied Humanities AI leadership - Ethan Mollick - One Useful Thing Early Childhood and developers
In this episode of Hashtag Trending, host Jim Love discusses NVIDIA CEO Jensen Huang's criticism of US export controls on AI chips that have led to significant financial losses for his company while bolstering Chinese AI competitors like Huawei. NVIDIA faces an $8 billion revenue loss due to restricted H20 chip exports to China. Huang argues that these policies are accelerating Chinese innovation and undermining US global leadership in AI technology. The episode also highlights Getty Images CEO Craig Peters' struggle with the high costs of litigating AI copyright infringement cases. Peters reveals that even a major company like Getty cannot afford to fight every instance of AI firms using copyrighted content without permission, creating a severe economic imbalance. The script ends with an exploration of the high rate of 'hallucinations' by AI in legal research and the resulting professional risks for lawyers, emphasizing the need for more stringent fact-checking. 00:00 Introduction and Headlines 00:26 NVIDIA's Struggles with US Export Controls 03:52 Getty Images' Battle Against AI Copyright Infringement 07:13 Legal Challenges with AI-Generated Fake Case Law 10:53 The Importance of Fact-Checking in AI Research 12:29 Conclusion and Viewer Engagement
Join Pathmonk Presents as we explore Deep Media with Ryan Ofman, Head of AI Research. Deep Media pioneers AI-driven deepfake detection, protecting government, social media, and financial sectors from misinformation. Ryan shares how they engage clients through conferences and viral deepfake detection reports, leveraging their website for education and acquisition. Learn about urgent “earthquake” conversions versus proactive “aftershock” inquiries, the importance of accessible AI explanations, and their work with academics for equitable AI. Discover tips for creating converting websites with engaging demos and compelling content. Tune in for insights on combating digital threats!
Zihan Wang is an AI researcher at Northwestern University, where he works on vision-language models, robotics, and reinforcement learning. Previously, he interned at DeepSeek, contributing to projects like DeepSeek-V2.Zihan's homepage: https://zihanwang314.github.io/(00:00) - Introduction (01:13) - Zihan's Background, CS and AI Research in China (11:09) - DeepSeek; Human capital flow from PRC to US (16:07) - DeepSeek, Open Source and AI Research (31:52) - Model Size and Performance Constraints (33:01) - Data Bottleneck in Pre-trained Models (34:12) - Transformer Architecture and Scaling Laws (36:30) - Efficiency in Model Training (47:44) - Chain of Experts Architecture (01:01:06) - Future of AI and Robotics Music used with permission from Blade Runner Blues Livestream improvisation by State Azure.–Steve Hsu is Professor of Theoretical Physics and of Computational Mathematics, Science, and Engineering at Michigan State University. Previously, he was Senior Vice President for Research and Innovation at MSU and Director of the Institute of Theoretical Science at the University of Oregon. Hsu is a startup founder (SuperFocus.ai, SafeWeb, Genomic Prediction, Othram) and advisor to venture capital and other investment firms. He was educated at Caltech and Berkeley, was a Harvard Junior Fellow, and has held faculty positions at Yale, the University of Oregon, and MSU. Please send any questions or suggestions to manifold1podcast@gmail.com or Steve on X @hsu_steve.
Yaron Singer, Vice President of AI and Security at Cisco, co-founded a company specializing in artificial intelligence solutions, which was acquired by Cisco in 2024. They developed a firewall for artificial intelligence, a tool designed to protect AI from making critical mistakes. No matter how sophisticated AI is, errors can still happen, and these errors can have far-reaching consequences. The product is designed to detect and fix such mistakes. This technology was developed long before ChatGPT and its competitors burst onto the scene, making it the hottest industry in tech investment. Join Singer as he sits down with UC San Diego professor Mikhail Belkin to discuss his work and the continued effort to make artificial intelligence secure. Series: "Data Science Channel" [Science] [Show ID: 40265]
Yaron Singer, Vice President of AI and Security at Cisco, co-founded a company specializing in artificial intelligence solutions, which was acquired by Cisco in 2024. They developed a firewall for artificial intelligence, a tool designed to protect AI from making critical mistakes. No matter how sophisticated AI is, errors can still happen, and these errors can have far-reaching consequences. The product is designed to detect and fix such mistakes. This technology was developed long before ChatGPT and its competitors burst onto the scene, making it the hottest industry in tech investment. Join Singer as he sits down with UC San Diego professor Mikhail Belkin to discuss his work and the continued effort to make artificial intelligence secure. Series: "Data Science Channel" [Science] [Show ID: 40265]
Yaron Singer, Vice President of AI and Security at Cisco, co-founded a company specializing in artificial intelligence solutions, which was acquired by Cisco in 2024. They developed a firewall for artificial intelligence, a tool designed to protect AI from making critical mistakes. No matter how sophisticated AI is, errors can still happen, and these errors can have far-reaching consequences. The product is designed to detect and fix such mistakes. This technology was developed long before ChatGPT and its competitors burst onto the scene, making it the hottest industry in tech investment. Join Singer as he sits down with UC San Diego professor Mikhail Belkin to discuss his work and the continued effort to make artificial intelligence secure. Series: "Data Science Channel" [Science] [Show ID: 40265]
Yaron Singer, Vice President of AI and Security at Cisco, co-founded a company specializing in artificial intelligence solutions, which was acquired by Cisco in 2024. They developed a firewall for artificial intelligence, a tool designed to protect AI from making critical mistakes. No matter how sophisticated AI is, errors can still happen, and these errors can have far-reaching consequences. The product is designed to detect and fix such mistakes. This technology was developed long before ChatGPT and its competitors burst onto the scene, making it the hottest industry in tech investment. Join Singer as he sits down with UC San Diego professor Mikhail Belkin to discuss his work and the continued effort to make artificial intelligence secure. Series: "Data Science Channel" [Science] [Show ID: 40265]
In this episode of The Persistent Entrepreneur, David Hill sits down with Amarpreet Kalkat, founder and CEO of Humantic AI, to explore how artificial intelligence is revolutionizing the sales process—before the first call even happens. Amarpreet shares how his platform uses AI to generate powerful personality insights and buyer intelligence that help sales professionals personalize outreach, build trust faster, and dramatically improve conversion rates. If you want to learn how to leverage AI to connect smarter and sell more effectively, this is a conversation you don't want to miss. Full Name: Amarpreet Kalkat Email: kalkat@humantic.ai Phone Number: +915103299004 Social Media Links: https://www.linkedin.com/in/amarpreetkalkat/ http://x.com/amarpreetkalkat https://kalkat.substack.com/ Connect with David LINKS: www.davidhill.ai SOCIALS: Facebook: https://www.facebook.com/davidihill/ LinkedIn: https://www.linkedin.com/in/davidihill YouTube: https://www.youtube.com/c/DavidHillcoach TikTok: www.tiktok.com/@davidihill Instagram: https://www.instagram.com/davidihill X: https://twitter.com/davidihill RING LEADER AI DEMO CALL- 774-214-2076 PODCAST SUBSCRIBE & REVIEW https://podcasts.apple.com/us/podcast/the-persistent-entrepreneur/id1081069895
Top Story: Visa trying to the AI shopping experience credit card AI continues to grow, but hallucinations are getting worse Zuckerberg thinks most of your friends will be AI chat bots Researchers violated ethical standards while investigating AI usage online
This and all episodes at: https://aiandyou.net/ . How to manage the integration of AI at scale into the enterprise is the territory of today's guest, Diane Gutiw, Vice President and leader of the AI research center at the global business consultancy CGI. She holds a PhD in Medical Information Technology Management and has led collaborative strategy design and implementation planning for advanced analytics and AI for large organizations in the energy and utilities, railway, and government healthcare sectors. In part 2, we talk about synthetic data, digital triplets, agentic AI and continuous autonomous improvement, and best practices for compliance. All this plus our usual look at today's AI headlines. Transcript and URLs referenced at HumanCusp Blog.
AI isn't just about flashy tools—it's about transforming how AEC firms research, analyze, and make decisions. In this episode of KP Unpacked - the number one Podcast in AEC, Jeff Echols and Frank Lazzaro dive deep into the power of AI research tools and how you can use them to work smarter, not harder.Key takeaways from this episode:Why AI research tools are changing the game for AEC firmsFrank's personal formula for finding 12 minutes of efficiency daily using AIReal-world examples of how AI tools outperform traditional research methodsThe top AI tools you should be using: ChatGPT Deep Research, Storm, and PerplexityHow to avoid AI's common pitfalls in data analysis and research workflowsActionable tips to start using AI research tools in your business todayWe want to tell you about something we've been quietly working on—and now it's live. It's called Catalyst. It's a space built for AEC professionals like you who are designing, building, and reimagining the future. Catalyst is where the sharpest minds in our industry connect, share, and lead what's next. If this sparked something, Catalyst is where we keep the conversation going.This uniquely active space is where the top minds in our industry come together and shape what's next in AEC. Ignite innovation. Join the waitlist at kpreddy.co. Hope to see you there!
Learn how to leverage advanced AI strategies that most business owners miss The untapped potential of artificial intelligence for business growth. The conversation explores how most people are barely scratching the surface of AI's capabilities, sharing practical examples of how AI can research prospects, create marketing campaigns, build prototypes, and even write books in a fraction of the time traditional methods require. Mike Koenigs is the author of "The AI Accelerator" and founder of the Superpower Accelerator. As one of Charles' early mentors, Mike has consistently stayed ahead of marketing trends and now helps business owners harness AI to expand their capabilities, create one-person marketing teams, elevate their authority, and scale without adding headcount. His innovative "thousand dollars cup of coffee" campaign demonstrates how AI-powered research can transform client consultations and dramatically increase revenue. KEY TAKEAWAYS: Most businesses only scratch the surface of AI potential, using basic ChatGPT instead of the full AI ecosystem. Quality AI output directly reflects quality input—effective prompt crafting is essential. Successful AI adoption requires understanding possibilities, selecting right tools, and focusing on four key functions. Being "tool agnostic" is crucial—multiple AI models provide better insights than any single tool. AI's greatest advantage comes from unleashing imagination and asking better questions. Create AI style guides to capture brand voice and apply consistently across all content. The "one-person marketing department" is now possible through comprehensive AI tools. Mike's "thousand dollars cup of coffee" model demonstrates how AI research transforms high-ticket sales. Websites: Main Website: https://superpoweraccelerator.com/ AI Accelerator: https://ai.mikekoenigs.com/ Book Website: https://aiaccelerator.mikekoenigs.com/ Social Media: LinkedIn: https://www.linkedin.com/in/mikekoenigs/ Growing your business is hard, but it doesn't have to be. In this podcast, we will be discussing top level strategies for both growing and expanding your business beyond seven figures. The show will feature a mix of pure content and expert interviews to present key concepts and fundamental topics in a variety of different formats. We believe that this format will enable our listeners to learn the most from the show, implement more in their businesses, and get real value out of the podcast. Enjoy the show. Please remember to rate, review and subscribe to the podcast so you don't miss any future episodes. Your support and reviews are important and help us to grow and improve the show. Follow Charles Gaudet and Predictable Profits on Social Media: Facebook: facebook.com/PredictableProfits Instagram: instagram.com/predictableprofits Twitter: twitter.com/charlesgaudet LinkedIn: linkedin.com/in/charlesgaudet Visit Charles Gaudet's Wesbites: www.PredictableProfits.com
Crafting Cannabis Episode #73 - BlueRidgeGrows is back! We got an update on the Helene aftermath, talked about sheep shearing and breeding autoflowers, and dove deep into AI research with cannabis health data, plus a lot more. Don't miss this one!----------------------------------Use code "craftingcannabispodcast" for 7% off at https://grovebags.com/Website: https://craftingcannabispodcast.comDiscord: https://discord.gg/craftingcannabispodcastInstagram: @craftingcannabispodcast @earlybird_autogrows @blackwatergrows @ruderalister
In this episode, Dr. Linda Chu speaks with Sarah Atzen, Lead Scientific Editor for Radiology, about best practices for writing AI research papers. They explore key tips from the recent article “Top 10 Tips for Writing about AI in Radiology” to help authors improve clarity, accuracy, and impact. Top 10 Tips for Writing about AI in Radiology: A Brief Guide for Authors. Atzen. Radiology 2025; 314(2):e243347.
This and all episodes at: https://aiandyou.net/ . How to manage the integration of AI at scale into the enterprise is the territory of today's guest, Diane Gutiw, Vice President and leader of the AI research center at the global business consultancy CGI. She holds a PhD in Medical Information Technology Management and has led collaborative strategy design and implementation planning for advanced analytics and AI for large organizations in the energy and utilities, railway, and government healthcare sectors. We talk about how enterprises manage the integration of AI at the dizzying speeds of change today, where AI does and does not impact employment, how the HR department should change in those enterprises, how to deal with hallucinations, and how to manage the risks of deploying generative AI in customer solutions. All this plus our usual look at today's AI headlines. Transcript and URLs referenced at HumanCusp Blog.
Are your mental health issues trapped between therapy appointments? When I hit a rough patch, I turned to AI for therapy out of desperation, and was genuinely shocked by the results. The AI asked me introspective questions identical to what my human therapist uses, but went further by offering specific suggestions my therapist couldn't provide. In this episode, we explore the surprising benefits and legitimate concerns of using AI for mental wellness. You'll discover how these tools can organize your scattered thoughts, provide actionable steps for improvement, and serve as a valuable bridge between professional therapy sessions. Listen now to learn how AI might become your unexpected mental health ally – just don't forget the human connection that algorithms can't replace. Topics Discussed: My personal experience using AI as a therapy tool How AI can provide structured introspection and actionable steps for mental health The surprising effectiveness of AI-generated journaling prompts for emotional connection The potential dangers of relying solely on AI for medical or mental health advice Privacy concerns and data collection risks when sharing personal information with AI Research comparing AI vs human doctor diagnostic accuracy (77% vs 67%) How AI lacks human imagination but excels at pattern recognition and memory recall The future integration of AI tools in professional healthcare settings Finding the balance between AI assistance and human connection Using AI to transform traditional journaling practices for deeper emotional connection ---- MORE FROM THE FIT MESS: Connect with us on Threads, Twitter, Instagram, Facebook, and Tiktok Subscribe to The Fit Mess on Youtube Join our community in the Fit Mess Facebook group ---- LINKS TO OUR PARTNERS: Take control of how you'd like to feel with Apollo Neuro Explore the many benefits of cold therapy for your body with Nurecover Muse's Brain Sensing Headbands Improve Your Meditation Practice. Get started as a Certified Professional Life Coach! Get a Free One Year Supply of AG1 Vitamin D3+K2, 5 Travel Packs Revamp your life with Bulletproof Coffee You Need a Budget helps you quickly get out of debt, and save money faster! Start your own podcast!
In this thought-provoking episode of Play the King Win the Day, host Brad Banyas sits down with Noah Kenney, (AI) Research & Development, ethical AI strategist, entrepreneur and founder of Disruptive AI Lab. Together, they explore the transformative impact of artificial intelligence (AI) across major industries—including healthcare, finance, and transportation.Noah Kenney shares expert insights on the future of autonomous vehicles, the role of AI in medical diagnostics, and the growing challenge of misinformation and algorithmic bias. The discussion dives into the ethical complexities of AI-generated content, intellectual property in the age of generative tools, and the global push for standardized, human-centered AI governance.This episode also introduces the Global Artificial Intelligence Framework (GAIF)—a pioneering set of guidelines designed to ensure AI is developed and deployed responsibly.Noah Kenney offers fresh insights into AI that will broaden your perspective—whether you're in technology, business, policy, or simply curious. Play the King, Win the Day!*wisdom to power your success.
AI Unraveled: Latest AI News & Trends, Master GPT, Gemini, Generative AI, LLMs, Prompting, GPT Store
Latest News on April 24th 2025 show OpenAI releasing a powerful image generation API adopted by major creative platforms and forecasting substantial future revenue. Microsoft unveiled AI agents for workplace collaboration, envisioning a future of human-AI partnerships. Anthropic's Claude introduced features to automate meeting preparation, enhancing productivity. Simultaneously, ethical considerations arose as former OpenAI staff challenged its shift to a for-profit model. In transportation, Tesla began supervised robotaxi testing, while in the chatbot market, Google's Gemini reported significant user growth, still trailing competitors.Want More AI Insights? Tune in to the AI Unraveled Podcast – your daily source for breaking news, expert interviews, and in-depth discussions on all things artificial intelligence at https://podcasts.apple.com/ca/podcast/ai-unraveled-latest-ai-news-trends-chatgpt-gemini-deepseek/id1684415169 Stay ahead of the curve in just a few minutes a day!Djamgatech PRO: Ad-Free Certification Mastery Unlock 100+ Professional Certifications with Premium Features
In this episode, Dipendra Kumar, Staff Research Scientist, and Alnur Ali, Staff Software Engineer at Databricks, discuss the challenges of applying AI in enterprise environments and the tools being developed to bridge the gap between research and real-world deployment.Highlights include:- The challenges of real-world AI—messy data, security, and scalability.- Why enterprises need high-accuracy, fine-tuned models over generic AI APIs.- How QuickFix learns from user edits to improve AI-driven coding assistance.- The collaboration between research & engineering in building AI-powered tools.- The evolving role of developers in the age of generative AI.
Episode: Balancing AI and Human Connection in B2B MarketingIn this episode of the B2B Marketing Excellence Podcast, I'm diving into a topic that's close to my heart — how we can embrace AI while still preserving the human touch that builds strong, lasting business relationships.At World Innovators, our family-run agency has spent the last 44 years helping B2B companies reach the right audience through strategic marketing rooted in trust, personalization, and genuine care. Today, AI is offering incredible tools to support those efforts — but it's how we use them that matters most.In this episode, I share real stories (including a powerful post-conference message I received) and practical ways to use AI as a supportive partner — not a substitute. Whether it's using AI for note-taking, preparing for outreach, or personalizing communication, I walk through how to make these tools work for your brand while keeping your values front and center.Let's talk about how to build deeper, more meaningful connections — with a little help from technology, and a lot of heart.⏱️ Episode Breakdown:• 00:00 – Introduction: Balancing AI and Human Connection• 00:13 – The Legacy of Building Relationships• 00:53 – Leveraging AI Without Losing the Human Touch• 02:05 – Real-Life Experiences and Insights• 02:57 – The Role of AI in Enhancing Relationships• 05:47 – Practical Applications of AI in Business• 08:40 – The Human Element in AI-Assisted Communication• 12:31 – AI as a Support System, Not a Substitute• 16:25 – Conclusion: Embracing AI for Deeper ConnectionsAt World Innovators, we believe B2B marketing should be about building relationships, not just generating clicks. For 44 years, we've helped industrial and executive education brands find the right people through trusted media sources, curated email lists, and strategic outreach.
In this episode, Josiah Mackenzie shares some top takeaways from his latest research, including how 87% of hospitality professionals participating in the study already use AI to improve efficiency, creativity, and guest experience. Listen now for practical examples, underutilized AI opportunities, and actionable insights you can use in your hotel or hospitality business.Also see:AI 2027 ProjectWhat AI Might Bring Hotels in 2025 - Martin SolerAmerica's Chief AI Officer for Travel Shares Advice for 2025 - Janette Roush, Brand USALess Ringing, More Hospitality: AI-Powered PBX To Give Our Teams More Time for Guests - Steven Marais, Noble House Hotels & ResortsAI & Hotel Tech Bets For Our People-First Approach - Dina Belon, Staypineapple HotelsThe Future of Hotel Management: Automation, AI, and Innovation - Sloan Dean, Remington HospitalityAI's Impact On Our Business - Ernest Lee, citizenMHow AI Helps Me Run More Profitable Hotels - Sean Murphy, The Bower50 Days, 50 Concepts: Rethinking Experiential Hospitality with Generative AI - Dylan Barahona A few more resources: If you're new to Hospitality Daily, start here. You can send me a message here with questions, comments, or guest suggestions If you want to get my summary and actionable insights from each episode delivered to your inbox each day, subscribe here for free. Follow Hospitality Daily and join the conversation on YouTube, LinkedIn, and Instagram. If you want to advertise on Hospitality Daily, here are the ways we can work together. If you found this episode interesting or helpful, send it to someone on your team so you can turn the ideas into action and benefit your business and the people you serve! Music for this show is produced by Clay Bassford of Bespoke Sound: Music Identity Design for Hospitality Brands
Alejandro and Julia of theluddite.org join us to debunk some terrible AI research, and the bad reporting compounding the problems on top of that. Also, what is AI? Can it ever think for itself? Are you an expert in something and want to be on the show? Apply here! Please support the show on patreon! You get ad free episodes, early episodes, and other bonus content! This content is CAN credentialed, which means you can report instances of harassment, abuse, or other harm on their hotline at (617) 249-4255, or on their website at creatoraccountabilitynetwork.org.
Members of Elon Musk's Department of Government Efficiency now have access to technical systems maintained by United States Citizenship and Immigration Services, according to a recent memorandum viewed by FedScoop. The memo, which was sent from and digitally signed by USCIS Chief Information Officer William McElhaney, states that Kyle Shutt, Edward Coristine, Aram Mogahaddassi and Payton Rehling were granted access to USCIS systems and data repositories, and that a Department of Homeland Security review was required to determine whether that access should continue. Coristine, 19, is one of the more polarizing members of DOGE. He previously provided assistance to a cybercrime ring through a company he operated while he was in high school, according to other news outlets. Coristine worked for a short period at Neuralink, Musk's brain implant company, and was previously stationed by DOGE at the Cybersecurity and Infrastructure Security Agency. The memo, dated March 28, asks DHS Deputy Secretary Troy Edgar to have his office review and provide direction for the four DOGE men regarding their access to the agency's “data lake” — called USCIS Data Business Intelligence Services — as well as two associated enabling technologies, Databricks and Github. The document says DHS CIO Antoine McCord and Michael Weissman, the agency's chief data officer, asked USCIS to enable Shutt and Coristine's access to the USCIS data lake in mid-March, and Mogahaddassi requested similar access days later. A bipartisan bill to fully establish a National Science Foundation-based resource aimed at providing essential tools for AI research to academics, nonprofits, small businesses and others was reintroduced in the House last week. Under the Creating Resources for Every American To Experiment with Artificial Intelligence (CREATE AI) Act of 2025 (H.R. 2385), a full-scale National AI Research Resource would be codified at NSF. While that resource currently exists in pilot form, legislation authorizing the NAIRR is needed to continue that work. Rep. Jay Obernolte, R-Calif., who sponsors the bill, said in a written statement announcing the reintroduction: “By empowering students, universities, startups, and small businesses to participate in the future of AI, we can drive innovation, strengthen our workforce, and ensure that American leadership in this critical field is broad-based and secure.” The NAIRR pilot, as it stands, is a collection of resources from the public and private sectors — such as computing power, storage, AI models, and data — that are made available to those researching AI to make the process of accessing those types of tools easier. The Daily Scoop Podcast is available every Monday-Friday afternoon. If you want to hear more of the latest from Washington, subscribe to The Daily Scoop Podcast on Apple Podcasts, Soundcloud, Spotify and YouTube.
OpenAI researcher Adam Kalai sits down with UC San Diego professor to discuss his work in machine learning, algorithmic fairness, and artificial intelligence. Kalai has contributed research in areas like fairness in AI models, word embeddings, and human-AI collaboration. He has worked at Microsoft Research and has published influential papers on bias in machine learning models. His work has helped shape discussions on ethical AI and the development of more equitable AI systems. Series: "Data Science Channel" [Science] [Show ID: 40264]
OpenAI researcher Adam Kalai sits down with UC San Diego professor Mikhail Belkin to discuss his work in machine learning, algorithmic fairness, and artificial intelligence. Kalai has contributed research in areas like fairness in AI models, word embeddings, and human-AI collaboration. He has worked at Microsoft Research and has published influential papers on bias in machine learning models. His work has helped shape discussions on ethical AI and the development of more equitable AI systems. Series: "Data Science Channel" [Science] [Show ID: 40264]
OpenAI researcher Adam Kalai sits down with UC San Diego professor Mikhail Belkin to discuss his work in machine learning, algorithmic fairness, and artificial intelligence. Kalai has contributed research in areas like fairness in AI models, word embeddings, and human-AI collaboration. He has worked at Microsoft Research and has published influential papers on bias in machine learning models. His work has helped shape discussions on ethical AI and the development of more equitable AI systems. Series: "Data Science Channel" [Science] [Show ID: 40264]
OpenAI researcher Adam Kalai sits down with UC San Diego professor Mikhail Belkin to discuss his work in machine learning, algorithmic fairness, and artificial intelligence. Kalai has contributed research in areas like fairness in AI models, word embeddings, and human-AI collaboration. He has worked at Microsoft Research and has published influential papers on bias in machine learning models. His work has helped shape discussions on ethical AI and the development of more equitable AI systems. Series: "Data Science Channel" [Science] [Show ID: 40264]
In today's show: The news of the day; Czech family's plea for bone marrow donor sparks nationwide response; bringing global and space phenomena to life in Žatec; and for our feature, we have Jakub Ferenčík's interview with Lea-Ann Germinder an AI researcher on Czech-US alignment in AI regulation, and more.
Send Everyday AI and Jordan a text messageStartups are changing quickly. That means Venture Capital is changing just as fast.
In this episode of the Security Matters podcast, host David Puner is joined by Lavi Lazarovitz, Vice President of Cyber Research at CyberArk Labs, to explore the transformative impact of AI agents on cybersecurity and automation. They discuss real-world scenarios where AI agents monitor security logs, flag anomalies, and automate responses, highlighting both the opportunities and risks associated with these advanced technologies.Lavi shares insights into the evolution of AI agents, from chatbots to agentic AI, and the challenges of building trust and resilience in AI-driven systems. The conversation delves into the latest research areas, including safety, privacy, and security, and examines how different industries are adopting AI agents to handle vast amounts of data.Tune in to learn about the critical security challenges posed by AI agents, the importance of trust in automation, and the strategies organizations can implement to protect their systems and data. Whether you're a cybersecurity professional or simply curious about the future of AI, this episode offers valuable insights into the rapidly evolving world of AI agents.
AI is changing the way businesses grow, and if you're not using it for research, you're falling behind. Whether it's market trends, audience insights, or competitor analysis, AI tools like ChatGPT and Gemini can give you a massive edge. In this podcast, I'll show you how to use AI research to make smarter business decisions, build a strong brand, and scale faster than ever.Want to see exactly how it works? Watch now and start leveraging AI to boost your business today.
Kevin Reid-Morris has led much of MDM's qualitative AI industry research over the past few years, and he's set to share the results of his major two-year project at our SHIFT conference. Here, he and Tom Gale discuss the transformative impact of AI in distribution and that project which identified over 50 practical AI use cases that distributors can implement to enhance operational efficiency and gain a competitive edge.
This episode is sponsored by Netsuite by Oracle, the number one cloud financial system, streamlining accounting, financial management, inventory, HR, and more. NetSuite is offering a one-of-a-kind flexible financing program. Head to https://netsuite.com/EYEONAI to know more. Can AI learn like humans? In this episode, Patrick Pilarski, Canada CIFAR AI Chair and professor at the University of Alberta, breaks down The Alberta Plan—a bold roadmap for achieving Artificial General Intelligence (AGI) through reinforcement learning and real-time experience-based AI. Unlike large pre-trained models that rely on massive datasets, The Alberta Plan champions continual learning, where AI evolves from raw sensory experience, much like a child learning through trial and error. Could this be the key to unlocking true intelligence? Pilarski also shares insights from his groundbreaking work in bionic medicine, where AI-powered prosthetics are transforming human-machine interaction. From neuroprostheses to reinforcement learning-driven robotics, this conversation explores how AI can enhance—not just replace—human intelligence. What You'll Learn in This Episode: Why reinforcement learning is a better path to AGI than pre-trained models The four core principles of The Alberta Plan and why they matter How AI-driven bionic prosthetics are revolutionizing human-machine integration The battle between reinforcement learning and traditional control systems in robotics Why continual learning is critical for AI to avoid catastrophic forgetting How reinforcement learning is already powering real-world breakthroughs in plasma control, industrial automation, and beyond The future of AI isn't just about more data—it's about AI that thinks, adapts, and learns from experience. If you're curious about the next frontier of AI, the rise of reinforcement learning, and the quest for true intelligence, this episode is a must-watch. Subscribe for more AI deep dives! (00:00) The Alberta Plan: A Roadmap to AGI (02:22) Introducing Patrick Pilarski (05:49) Breaking Down The Alberta Plan's Core Principles (07:46) The Role of Experience-Based Learning in AI (08:40) Reinforcement Learning vs. Pre-Trained Models (12:45) The Relationship Between AI, the Environment, and Learning (16:23) The Power of Reward in AI Decision-Making (18:26) Continual Learning & Avoiding Catastrophic Forgetting (21:57) AI in the Real World: Applications in Fusion, Data Centers & Robotics (27:56) AI Learning Like Humans: The Role of Predictive Models (31:24) Can AI Learn Without Massive Pre-Trained Models? (35:19) Control Theory vs. Reinforcement Learning in Robotics (40:16) The Future of Continual Learning in AI (44:33) Reinforcement Learning in Prosthetics: AI & Human Interaction (50:47) The End Goal of The Alberta Plan
In this episode of “Waking Up With AI,” Katherine Forrest and Anna Gressel examine the integration of end-to-end reasoning and agentic AI capabilities, with new developments from OpenAI, DeepMind and other leading AI labs. Katherine also shares her firsthand experience with OpenAI's new deep research capability, which is transforming academic applications of AI. ## Learn More About Paul, Weiss's Artificial Intelligence Practice: https://www.paulweiss.com/practices/litigation/artificial-intelligence
Most people are barely scratching the surface of what generative AI can do. While some fear it will replace their jobs, others dismiss it as a passing trend—but both extremes miss the point. In this episode, Ashok Sivanand breaks down the real opportunity AI presents: not as a replacement for human judgment, but as a powerful tool that can act as both a dutiful intern and an expert consultant. Learn how to integrate AI into your daily work, from automating tedious tasks to sharpening your strategic thinking, all while staying in control. Unlock the full potential of your product team with Integral's player coaches, experts in lean, human-centered design. Visit integral.io/convergence for a free Product Success Lab workshop to gain clarity and confidence in tackling any product design or engineering challenge. Inside the episode... Why so few people are using generative AI daily—and why that needs to change The two key roles AI can play: the intern and the consultant How AI can help professionals streamline research, analysis, and decision-making Practical prompts and frameworks for getting the most out of AI tools The dangers of "AI autopilot" and why staying in the driver's seat is critical Security and privacy concerns: What every AI user should know The best AI tools for different use cases—beyond just ChatGPT How companies can encourage AI adoption without creating unnecessary friction Mentioned in this episode AI Tools: ChatGPT, Claude, Perplexity, Gemini, Copilot, Grok Amazon's six-page memo template for effective decision-making: https://medium.com/@info_14390/the-ultimate-guide-to-amazons-6-pager-memo-method-c4b683441593 Ready Signal for external market factor analysis: https://www.readysignal.com/ AI prompting frameworks from Geoff Woods of AI Leadership: https://www.youtube.com/watch?v=HToY8gDTk6E Andrej Karpathy's Deep Dive into LLMs: https://www.youtube.com/watch?v=7xTGNNLPyMI Books by Carmine Gallo: The Presentation Secrets of Steve Jobs & Talk Like TED: https://www.amazon.com/Presentation-Secrets-Steve-Jobs-Insanely/dp/1491514310 Subscribe to the Convergence podcast wherever you get podcasts—including video episodes on YouTube at youtube.com/@convergencefmpodcast Learn something? Give the podcast a 5-star review and like the episode on YouTube. It's how the show grows. Follow the Pod Linkedin: https://www.linkedin.com/company/convergence-podcast/ X: https://twitter.com/podconvergence Instagram: @podconvergence
Artificial intelligence is radically transforming software development. AI-assisted coding tools are generating billions in investment, promising faster development cycles, and shifting engineering roles from code authors to code editors. But how does this impact software quality, security, and team dynamics? How can product teams embrace AI without falling into the hype? In this episode, AI assisted Agile expert Mike Gehard shares his hands-on experiments with AI in software development. From his deep background at Pivotal Labs to his current work pushing the boundaries of AI-assisted coding, Mike reveals how AI tools can amplify quality practices, speed up prototyping, and even challenge the way we think about source code. He discusses the future of pair programming, the evolving role of test-driven development, and how engineers can better focus on delivering user value. Unlock the full potential of your product team with Integral's player coaches, experts in lean, human-centered design. Visit integral.io/convergence for a free Product Success Lab workshop to gain clarity and confidence in tackling any product design or engineering challenge. Inside the episode... Mike's background at Pivotal Labs and why he kept returning How AI is changing the way we think about source code as a liability Why test-driven development still matters in an AI-assisted world The future of pair programming with AI copilots The importance of designing better software in an AI-driven development process Using AI to prototype faster and build user-facing value sooner Lessons learned from real-world experiments with AI-driven development The risks of AI-assisted software, from hallucinations to security Mentioned in this episode Mike's Substack: https://aiassistedagiledevelopment.substack.com/ Mike's Github repo: https://github.com/mikegehard/ai-assisted-agile-development Pivotal Labs: https://en.wikipedia.org/wiki/Pivotal_Labs 12-Factor Apps: https://12factor.net/ GitHub Copilot: https://github.com/features/copilot Cloud Foundry: https://en.wikipedia.org/wiki/Cloud_Foundry Lean Startup by Eric Ries: https://www.amazon.com/Lean-Startup-Entrepreneurs-Continuous-Innovation/dp/0307887898 Refactoring by Martin Fowler and Kent Beck https://www.amazon.com/Refactoring-Improving-Existing-Addison-Wesley-Signature/dp/0134757599 Dependabot: https://github.com/dependabot Tessl CEO Guy Podjarny's talk: https://youtu.be/e1a3WuxTY-k Aider AI Pair programming terminal: https://aider.chat/ Gemini LLM: https://gemini.google.com/app Perplexity AI: https://www.perplexity.ai/ DeepSeek: https://www.deepseek.com/ Ian Cooper's talk on TDD: https://www.youtube.com/watch?v=IN9lftH0cJc Mike's newest Mountain Bike IBIS Ripmo V2S: https://www.ibiscycles.com/bikes/past-models/ripmo-v2s Mike's recommended house slippers: https://us.giesswein.com/collections/mens-wool-slippers/products/wool-slippers-dannheim Sorba Chattanooga Mountain Biking Trails: https://www.sorbachattanooga.org/localtrails Subscribe to the Convergence podcast wherever you get podcasts, including video episodes on YouTube at youtube.com/@convergencefmpodcast Learn something? Give us a 5-star review and like the podcast on YouTube. It's how we grow.
There's a huge number of AI tools emerging and we're testing them to see if they can help with different aspects of investing. From filtering, researching, and valuing opportunities to constructing a portfolio and monitoring positions - the impact of AI on investing is going to be profound.This year, we want to trial as many platforms as possible and share how we think about incorporating them into our process.In today's episode we trial Google's Notebook LM. Check out Ren's notes.—------Want to get involved in the podcast? Record a voice note or send us a message on our website and we'll play it on the podcast.—------Keep up with the news moving markets with the Equity Mates daily email and podcast:Sign up to our daily email to get the news delivered to your inbox at 6am every weekday morningPrefer to hear the news? We've turned our email into a podcast using AI - listen on Apple or Spotify—------Want more Equity Mates?Listen to our basics-of-investing podcast: Get Started Investing (Apple | Spotify)Watch Equity Mates on YouTubePick up our books: Get Started Investing and Don't Stress, Just InvestFollow us on social media: Instagram, TikTok, & LinkedIn—------In the spirit of reconciliation, Equity Mates Media and the hosts of Equity Mates Investing acknowledge the Traditional Custodians of country throughout Australia and their connections to land, sea and community. We pay our respects to their elders past and present and extend that respect to all Aboriginal and Torres Strait Islander people today. —------Equity Mates Investing is a product of Equity Mates Media. This podcast is intended for education and entertainment purposes. Any advice is general advice only, and has not taken into account your personal financial circumstances, needs or objectives. Before acting on general advice, you should consider if it is relevant to your needs and read the relevant Product Disclosure Statement. And if you are unsure, please speak to a financial professional. Equity Mates Media operates under Australian Financial Services Licence 540697. Hosted on Acast. See acast.com/privacy for more information.
It's that time of week where I'll take you through a rundown on some of the latest happenings at the critical intersection of business, tech, and human experience. While love is supposed to be in the air given it's Valentine's Day, I'm not sure the headlines got the memo.With that, let's get started.Elon's $97B OpenAI Takeover Stunt - Musk made a shock bid to buy OpenAI for $97 billion, raising questions about his true motives. Given his history with OpenAI and his own AI venture (xAI), this move had many wondering if he was serious or just trolling. Given OpenAI is hemorrhaging cash alongside its plans to pivot to a for-profit model, Altman is in a tricky position. Musk's bid seems designed to force OpenAI into staying a nonprofit, showing how billionaires use their wealth to manipulate industries, not always in ways that benefit the public.Is Google Now Pro-Harmful AI? - Google silently removed its long-standing ethical commitment to not creating AI for harmful purposes. This change, combined with its growing partnerships in military AI, raises major concerns about the direction big tech is taking. It's worth exploring how AI development is shifting toward militarization and how companies like Google are increasingly prioritizing government and defense contracts over consumer interests.The AI Agent Hype Cycle - AI agents are being hyped as the future of work, with companies slashing jobs in anticipation of AI taking over. However, there's more than meets the eye. While AI agents are getting more powerful, they're still unreliable, messy, and require human oversight. Companies are overinvesting in AI agents and quickly realizing they don't work as well as advertised. While that may sound good for human workers, I predict it will get worse before it gets better.Does Microsoft Research Show AI is Killing Critical Thinking? - A recent Microsoft study is making waves with claims that AI is eroding critical thinking and creativity. This week, I took a closer look at the research and explained why the media's fearmongering isn't entirely accurate. And yet, we should take this seriously. The real issue isn't AI itself; it's how we use it. If we keep becoming over-reliant on AI for thinking, problem-solving, and creativity, it will inevitably lead to cognitive atrophy.Show Notes:In this Weekly Update, Christopher explores the latest developments at the intersection of business, technology, and the human experience. The episode covers Elon Musk's surprising $97 billion bid to acquire OpenAI, its implications, and the debate over whether OpenAI should remain a nonprofit. The discussion also explores the military applications of AI, Google's recent shift away from its 'don't create harmful AI' policy, and the consequences of large-scale investments in AI for militaristic purposes. Additionally, Christopher examines the rise of AI agents, their potential to change the workforce, and the challenges they present. Finally, Microsoft's study on the erosion of critical thinking and empathy due to AI usage is analyzed, emphasizing the need for thoughtful and intentional application of AI technologies.00:00 - Introduction01:53 - Elon Musk's Shocking Offer to Buy OpenAI15:27 - Google's Controversial Shift in AI Ethics27:20 - Navigating the Hype of AI Agents29:41 - The Rise of AI Agents in the Workplace41:35 - Does AI Destroy Critical Thinking in Humans?52:49 - Concluding Thoughts and Future Outlook#AI #OpenAI #Microsoft #CriticalThinking #ElonMusk
AI copilots have changed a range of professions, from healthcare to finance, by automating tasks and enhancing productivity. But can copilots also create value for people performing more mechanical, hands-on tasks or figuring out how to bring factories online? In this episode, Barbara welcomes Olympia Brikis, Director of AI Research at Siemens, to show how generative AI is shaping new industrial tech jobs at the convergence of the real and digital worlds. Olympia sheds light on the unique career opportunities in AI and what it takes to thrive in this dynamic, emerging field. Whether you're a tech enthusiast or someone curious about tech careers, this episode offers a unique perspective on how AI is reshaping the landscape of mechanical and industrial professions. Tune in to learn about the exciting innovations and the future of AI in industry! Show notes In this episode, Barbara asks Olympia to share some resources that can help all of us get smarter on industrial AI. Here are Olympia's recommendations: For everyone just getting started with (Generative) AI: Elements of AI – great for learning how AI work and what it is https://www.elementsofai.com/ Generative AI for Everyone: https://www.coursera.org/learn/generative-ai-for-everyone Co-Intelligence: Living and Working with AI, by Ethan Mollick For those want to dive deeper into the technical aspects of Deep Neural Networks and Generative AI: Deep Learning Specialization: https://www.coursera.org/specializations/deep-learning Stanford University Lecture CS336: Language Modeling from Scratch https://stanford-cs336.github.io/spring2024/
Hey tech lovers! In this episode of The LEO Podcast, we dive into three hot tech stories. First, El Salvador is officially stepping away from Bitcoin as legal tender—was the crypto experiment a failure, or just ahead of its time? Then, AI-powered research assistants from OpenAI and Google are changing how scientists work, but can we really trust them to get the facts right? And finally, Apple is making big moves in Latin America, with eSIM technology giving it an edge over Samsung in the region. Tune in for all this and more onThe LEO Podcast!
This episode is sponsored by Thuma. Thuma is a modern design company that specializes in timeless home essentials that are mindfully made with premium materials and intentional details. To get $100 towards your first bed purchase, go to http://thuma.co/eyeonai In this episode of the Eye on AI podcast, Pedro Domingos, renowned AI researcher and author of The Master Algorithm, joins Craig Smith to explore the evolution of machine learning, the resurgence of Bayesian AI, and the future of artificial intelligence. Pedro unpacks the ongoing battle between Bayesian and Frequentist approaches, explaining why probability is one of the most misunderstood concepts in AI. He delves into Bayesian networks, their role in AI decision-making, and how they powered Google's ad system before deep learning. We also discuss how Bayesian learning is still outperforming humans in medical diagnosis, search & rescue, and predictive modeling, despite its computational challenges. The conversation shifts to deep learning's limitations, with Pedro revealing how neural networks might be just a disguised form of nearest-neighbor learning. He challenges conventional wisdom on AGI, AI regulation, and the scalability of deep learning, offering insights into why Bayesian reasoning and analogical learning might be the future of AI. We also dive into analogical learning—a field championed by Douglas Hofstadter—exploring its impact on pattern recognition, case-based reasoning, and support vector machines (SVMs). Pedro highlights how AI has cycled through different paradigms, from symbolic AI in the '80s to SVMs in the 2000s, and why the next big breakthrough may not come from neural networks at all. From theoretical AI debates to real-world applications, this episode offers a deep dive into the science behind AI learning methods, their limitations, and what's next for machine intelligence. Don't forget to like, subscribe, and hit the notification bell for more expert discussions on AI, technology, and the future of innovation! Stay Updated: Craig Smith Twitter: https://twitter.com/craigss Eye on A.I. Twitter: https://twitter.com/EyeOn_AI (00:00) Introduction (02:55) The Five Tribes of Machine Learning Explained (06:34) Bayesian vs. Frequentist: The Probability Debate (08:27) What is Bayes' Theorem & How AI Uses It (12:46) The Power & Limitations of Bayesian Networks (16:43) How Bayesian Inference Works in AI (18:56) The Rise & Fall of Bayesian Machine Learning (20:31) Bayesian AI in Medical Diagnosis & Search and Rescue (25:07) How Google Used Bayesian Networks for Ads (28:56) The Role of Uncertainty in AI Decision-Making (30:34) Why Bayesian Learning is Computationally Hard (34:18) Analogical Learning – The Overlooked AI Paradigm (38:09) Support Vector Machines vs. Neural Networks (41:29) How SVMs Once Dominated Machine Learning (45:30) The Future of AI – Bayesian, Neural, or Hybrid? (50:38) Where AI is Heading Next
The release of DeepSeek's AI models at the end of January 2025 sent shockwaves around the world. The weeks that followed have been rife with hype and rumor, ranging from suggestions that DeepSeek has completely upended the tech industry to claims the efficiency gains ostensibly unlocked by DeepSeek are exagerrated. So, what's the reality? And what does it all really mean for the tech industry? In this episode of the Technology Podcast, two of Thoughtworks' AI leaders — Prasanna Pendse (Global Director of AI Strategy) and Shayan Mohanty (Head of AI Research) — join hosts Prem Chandrasekaran and Ken Mugrage to provide a much-needed clear and sober perspective on DeepSeek. They dig into some of the technical details and discuss how the DeepSeek team was able to optimize the limited hardware at their disposal, and think through what the implications might be for the industry in the months to come. Read Prasanna's take on DeepSeek on the Thoughtworks blog: https://www.thoughtworks.com/insights/blog/generative-ai/demystifying-deepseek
Niloofar is a Postdoctoral researcher at University of Washington with research interests in building privacy preserving AI systems and studying the societal implications of machine learning models. She received her PhD in Computer Science from UC San Diego in 2023 and has received multiple awards and honors for research contributions. Time stamps of the conversation 00:00:00 Highlights 00:01:35 Introduction 00:02:56 Entry point in AI 00:06:50 Differential privacy in AI systems 00:11:08 Privacy leaks in large language models 00:15:30 Dangers of training AI on public data on internet 00:23:28 How auto-regressive training makes things worse 00:30:46 Impact of Synthetic data for fine-tuning 00:37:38 Most critical stage in AI pipeline to combat data leaks 00:44:20 Contextual Integrity 00:47:10 Are LLMs creative? 00:55:24 Under vs. Overpromises of LLMs 01:01:40 Publish vs. perish culture in AI research recently 01:07:50 Role of academia in LLM research 01:11:35 Choosing academia vs. industry 01:17:34 Mental Health and overarching More about Niloofar: https://homes.cs.washington.edu/~niloofar/ And references to some of the papers discussed: https://arxiv.org/pdf/2310.17884 https://arxiv.org/pdf/2410.17566 https://arxiv.org/abs/2202.05520 About the Host: Jay is a PhD student at Arizona State University working on improving AI for medical diagnosis and prognosis. Linkedin: https://www.linkedin.com/in/shahjay22/ Twitter: https://twitter.com/jaygshah22 Homepage: http://jayshah.me/ for any queries. Stay tuned for upcoming webinars! ***Disclaimer: The information in this video represents the views and opinions of the speaker and does not necessarily represent the views or opinions of any institution. It does not constitute an endorsement by any Institution or its affiliates of such video content.***
How can AI revolutionize the way we research and understand complex topics? In this Tech Talks Daily episode, I speak with Mel Morris, founder and CEO of Corpora.AI. This research engine redefines how individuals, businesses, and institutions approach knowledge discovery. With traditional search engines struggling to deliver depth and AI tools often relying on outdated or incomplete data, Corpora.AI takes a different approach. The platform processes millions of documents per second using advanced AI and proprietary language graph technology, delivering research reports with real-time insights and source attribution. Unlike conventional AI models that generate content from limited datasets, Corpora.AI dynamically ingests over 100 petabytes of open-source intelligence, ensuring users can access the most comprehensive, accurate, and up-to-date information. Mel shares his vision for democratizing access to high-level research, making it possible for users across academia, medicine, law, finance, government, and journalism to gain deeper insights faster. We explore how Corpora.AI's real-time data ingestion and multilingual capabilities allow professionals to conduct advanced research in one language and receive results in another. From patent research and market analysis to education and rapid learning, the applications of this research engine extend far beyond what traditional AI-powered search tools can offer. We also discuss how Corpora.AI is tackling some of the biggest challenges in AI-driven research, including bias, credibility, and transparency. By providing research reports with 400-500 cited sources per query, the platform ensures that every insight is traceable, allowing users to verify information and make informed decisions. With AI reshaping how we access and interpret knowledge, what does the future hold for research, education, and data-driven decision-making? Will AI-driven research engines like Corpora.AI replace traditional search methods? And how can businesses and institutions leverage these tools to stay ahead of the curve? Join me for this fascinating discussion as we explore the future of AI-powered research and how Corpora.AI is setting a new standard for knowledge discovery.
This episode is sponsored by Netsuite by Oracle, the number one cloud financial system, streamlining accounting, financial management, inventory, HR, and more. NetSuite is offering a one-of-a-kind flexible financing program. Head to https://netsuite.com/EYEONAI to know more. In this episode of the Eye on AI podcast, we dive into the transformative world of AI compute infrastructure with Mitesh Agrawal, Head of Cloud/COO at Lambda Mitesh takes us on a journey from Lambda Labs' early days as a style transfer app to its rise as a leader in providing scalable, deep learning infrastructure. Learn how Lambda Labs is reshaping AI compute by delivering cutting-edge GPU solutions and accessible cloud platforms tailored for developers, researchers, and enterprises alike. Throughout the episode, Mitesh unpacks Lambda Labs' unique approach to optimizing AI infrastructure—from reducing costs with transparent pricing to tackling the global GPU shortage through innovative supply chain strategies. He explains how the company supports deep learning workloads, including training and inference, and why their AI cloud is a game-changer for scaling next-gen applications. We also explore the broader landscape of AI, touching on the future of AI compute, the role of reasoning and video models, and the potential for localized data centers to meet the growing demand for low-latency solutions. Mitesh shares his vision for a world where AI applications, powered by Lambda Labs, drive innovation across industries. Tune in to discover how Lambda Labs is democratizing access to deep learning compute and paving the way for the future of AI infrastructure. Don't forget to like, subscribe, and hit the notification bell to stay updated on the latest in AI, deep learning, and transformative tech! Stay Updated: Craig Smith Twitter: https://twitter.com/craigss Eye on A.I. Twitter: https://twitter.com/EyeOn_AI (00:00) Introduction and Lambda Labs' Mission (01:37) Origins: From DreamScope to AI Compute Infrastructure (04:10) Pivoting to Deep Learning Infrastructure (06:23) Building Lambda Cloud: An AI-Focused Cloud Platform (09:16) Transparent Pricing vs. Hyperscalers (12:52) Managing GPU Supply and Demand (16:34) Evolution of AI Workloads: Training vs. Inference (20:02) Why Lambda Labs Sticks with NVIDIA GPUs (24:21) The Future of AI Compute: Localized Data Centers (28:30) Global Accessibility and Regulatory Challenges (32:13) China's AI Development and GPU Restrictions (39:50) Scaling Lambda Labs: Data Centers and Growth (45:22) Advancing AI Models and Video Generation (50:24) Optimism for AI's Future (53:48) How to Access Lambda Cloud
This week, we are joined by Andrew Morris, Founder and CTO of GreyNoise, to discuss their work on "GreyNoise Intelligence Discovers Zero-Day Vulnerabilities in Live Streaming Cameras with the Help of AI." GreyNoise discovered two critical zero-day vulnerabilities in IoT-connected live streaming cameras, used in sensitive environments like healthcare and industrial operations, by leveraging its AI-powered detection system, Sift. The vulnerabilities, CVE-2024-8956 (insufficient authentication) and CVE-2024-8957 (OS command injection), could allow attackers to take full control of affected devices, manipulate video feeds, or integrate them into botnets for broader attacks. This breakthrough underscores the transformative role of AI in identifying threats that traditional systems might miss, highlighting the urgent need for robust cybersecurity measures in the expanding IoT landscape. The research can be found here: GreyNoise Intelligence Discovers Zero-Day Vulnerabilities in Live Streaming Cameras with the Help of AI Learn more about your ad choices. Visit megaphone.fm/adchoices
This Week in Machine Learning & Artificial Intelligence (AI) Podcast
Today, we're joined by Arash Behboodi, director of engineering at Qualcomm AI Research to discuss the papers and workshops Qualcomm will be presenting at this year's NeurIPS conference. We dig into the challenges and opportunities presented by differentiable simulation in wireless systems, the sciences, and beyond. We also explore recent work that ties conformal prediction to information theory, yielding a novel approach to incorporating uncertainty quantification directly into machine learning models. Finally, we review several papers enabling the efficient use of LoRA (Low-Rank Adaptation) on mobile devices (Hollowed Net, ShiRA, FouRA). Arash also previews the demos Qualcomm will be hosting at NeurIPS, including new video editing diffusion and 3D content generation models running on-device, Qualcomm's AI Hub, and more! The complete show notes for this episode can be found at https://twimlai.com/go/711.