POPULARITY
Most brands are about to vanish from search. Yours doesn't have to.AI search isn't the future. It's already rewriting the rules.And if you're not adapting -- you're disappearing.What's changing? Who's winning?And why are some brands thriving while others fade into the algorithmic abyss?Chris Andrew, CEO & Co-founder of Scrunch AI, joins us to break it all down.Newsletter: Sign up for our free daily newsletterMore on this Episode: Episode PageJoin the discussion: Thoughts on this? Join the convo.Upcoming Episodes: Check out the upcoming Everyday AI Livestream lineupWebsite: YourEverydayAI.comEmail The Show: info@youreverydayai.comConnect with Jordan on LinkedInTopics Covered in This Episode:AI Search's Impact on Brand VisibilityStrategies for Winning in AI SearchAI Search and Customer Journey ChangesImportance of AI Crawlers in SEOShifting SEO Tactics for AI SearchAI and Third-Party Content InfluenceSmall Brands Competing in AI SearchFuture of All Search as AI SearchTimestamps:00:00 Brands in the age of AI search 02:50 Leveraging AI for Immediate Impact13:16 "Optimizing Content for AI Crawlers"15:55 "Unblocking AI Crawlers Essential"20:24 Rapid AI Developments Challenge Adaptation22:20 Optimizing Content for AI Retrieval24:31 AI Strategies for Online Brand Management28:22 ChatGPT Memory and AI PersonalizationKeywords:AI search, brand optimization, GPT, perplexity, customer journey, enterprise platform, AI crawlers, AI overview, Anthropic, Claude AI assistant, web research, deep research, Google Workspace, Microsoft Copilot, Google Gemini, VO two, AI video generator, text prompts, OpenAI, social network, CEO Sam Altman, AI-powered sharing, AI referral traffic, brand reputation, persona mapping, buyer behavior, ChatGPT, integration, Claude's new features, beta features, content strategy, organic search, content creation, user intent, AI monitoring, third party content, brand perception, intent-based content, personalized content, buyer intent, search behavior, buyer journey, market adaptation, business strategies, AI consumer, content optimization.Send Everyday AI and Jordan a text message. (We can't reply back unless you leave contact info) Try Google Veo 3 today! Sign up at gemini.google to get started. Try Google Veo 3 today! Sign up at gemini.google to get started.
Sander Schulhoff is the OG prompt engineer. He created the very first prompt engineering guide on the internet (two months before ChatGPT's release) and recently wrote the most comprehensive study of prompt engineering ever conducted (co-authored with OpenAI, Microsoft, Google, Princeton, and Stanford), analyzing over 1,500 academic papers and covering more than 200 prompting techniques. He also partners with OpenAI to run what was the first and is the largest AI red teaming competition, HackAPrompt, which helps discover the most state-of-the-art prompt injection techniques (i.e. ways to get LLMS to do things it shouldn't). Sander teaches AI red teaming on Maven, advises AI companies on security, and has educated millions of people on the most state-of-the-art prompt engineering techniques.In this episode, you'll learn:1. The 5 most effective prompt engineering techniques2. Why “role prompting” and threatening the AI no longer works—and what to do instead3. The two types of prompt engineering: conversational and product/system prompts4. A primer on prompt injection and AI red teaming—including real jailbreak tactics that are still fooling top models5. Why AI agents and robots will be the next major security threat6. How to get started in AI red teaming and prompt engineering7. Practical defense to put in place for your AI products—Brought to you by:Eppo—Run reliable, impactful experimentsStripe—Helping companies of all sizes grow revenueVanta—Automate compliance. Simplify security—Where to find Sander Schulhoff:• X: https://x.com/sanderschulhoff• LinkedIn: https://www.linkedin.com/in/sander-schulhoff/• Website: https://sanderschulhoff.com/• AI Red Teaming and AI Security Masterclass on Maven: https://bit.ly/44lLSbC• Free Lightning Lesson “How to Secure Your AI System” on 6/24: https://bit.ly/4ld9vZL—Where to find Lenny:• Newsletter: https://www.lennysnewsletter.com• X: https://twitter.com/lennysan• LinkedIn: https://www.linkedin.com/in/lennyrachitsky/—In this episode, we cover:(00:00) Introduction to Sander Schulhoff(04:29) The importance of prompt engineering(06:30) Real-world applications and examples(10:54) Basic prompt engineering techniques(23:46) Advanced prompt engineering techniques(29:00) The role of context and additional information(39:24) Ensembling techniques and thought generation(49:48) Conversational techniques for better results(50:46) Introduction to prompt injection(52:27) AI red teaming and competitions(54:23) The growing importance of AI security(01:02:45) Techniques to bypass AI safeguards(01:05:21) Challenges in AI security and future outlook(01:18:33) Misalignment and AI's potential risks(01:25:03) Final thoughts and lightning round—Referenced:• Reid Hoffman's tweet about using AI agents: https://x.com/reidhoffman/status/1930416063616884822• AI Engineer World's Fair: https://www.ai.engineer/• What Is Artificial Social Intelligence?: https://learnprompting.org/blog/asi• Devin: https://devin.ai/• Cursor: https://www.cursor.com/• Inside Devin: The world's first autonomous AI engineer that's set to write 50% of its company's code by end of year | Scott Wu (CEO and co-founder of Cognition): https://www.lennysnewsletter.com/p/inside-devin-scott-wu• The rise of Cursor: The $300M ARR AI tool that engineers can't stop using | Michael Truell (co-founder and CEO): https://www.lennysnewsletter.com/p/the-rise-of-cursor-michael-truell• Granola: https://www.granola.ai/• Building Lovable: $10M ARR in 60 days with 15 people | Anton Osika (CEO and co-founder): https://www.lennysnewsletter.com/p/building-lovable-anton-osika• Inside Bolt: From near-death to ~$40m ARR in 5 months—one of the fastest-growing products in history | Eric Simons (founder & CEO of StackBlitz): https://www.lennysnewsletter.com/p/inside-bolt-eric-simons• Behind the product: Replit | Amjad Masad (co-founder and CEO): https://www.lennysnewsletter.com/p/behind-the-product-replit-amjad-masad• Everyone's an engineer now: Inside v0's mission to create a hundred million builders | Guillermo Rauch (founder and CEO of Vercel, creators of v0 and Next.js): https://www.lennysnewsletter.com/p/everyones-an-engineer-now-guillermo-rauch• Technique #3: Examples in Prompts: From Zero-Shot to Few-Shot: https://learnprompting.org/docs/basics/few_shot?srsltid=AfmBOor2owyGXtzJZ8n0fJVCctM7UPZgZmH-mBuxRW4t9-kkaMd3LJVv• The Prompt Report: Insights from the Most Comprehensive Study of Prompting Ever Done: https://learnprompting.org/blog/the_prompt_report?srsltid=AfmBOoo7CRNNCtavzhyLbCMxc0LDmkSUakJ4P8XBaITbE6GXL1i2SvA0• State-of-the-Art Prompting for AI Agents | Y Combinator: https://www.youtube.com/watch?v=DL82mGde6wo• Use XML tags to structure your prompts: https://docs.anthropic.com/en/docs/build-with-claude/prompt-engineering/use-xml-tags• Role Prompting: https://learnprompting.org/docs/basics/roles?srsltid=AfmBOor2jcxJQvWBZyFa030Qt0fIIov3hSiWvI9VFyjO-Qp478EPJIU7• Is Role Prompting Effective?: https://learnprompting.org/blog/role_prompting?srsltid=AfmBOooiiyLD-0CsCYZ4m3SDhYOmtTyaTzeDo0FvK_i1x1gLM8MJS-Sn• Introduction to Decomposition Prompting Techniques: https://learnprompting.org/docs/advanced/decomposition/introduction?srsltid=AfmBOoojJmTQgBlmSlGYQ8kl-JPpVUlLKkL4YcFGS5u54JyeumUwlcBI• LLM Self-Evaluation: https://learnprompting.org/docs/reliability/lm_self_eval• Philip Resnik on X: https://x.com/psresnik• Anthropic's CPO on what comes next | Mike Krieger (co-founder of Instagram): https://www.lennysnewsletter.com/p/anthropics-cpo-heres-what-comes-next• Introduction to Ensembling Prompting: https://learnprompting.org/docs/advanced/ensembling/introduction?srsltid=AfmBOooGSyqsrjnEbXSYoKpG0ZlpT278NHQA6Fd8gMvNTJlWu7-qEYzh• Random forest: https://en.wikipedia.org/wiki/Random_forest• Chain-of-Thought Prompting: https://learnprompting.org/docs/intermediate/chain_of_thought?srsltid=AfmBOoqwE7SXlluy2sx_QY_VOKduyBplWtIWKEJaD6FkJW3TqeKPSJfx• Prompt Injecting: https://learnprompting.org/docs/prompt_hacking/injection?srsltid=AfmBOoqGgqbfXStrD6vlw5jy8HhEaESgGo2e57jyWL8lkZKktt_P6Zvn• Announcing HackAPrompt 2.0: The World's Largest AI Red-Teaming Hackathon: https://learnprompting.org/blog/announce-hackaprompt-2?srsltid=AfmBOopXKsHxy4aUtsvPCUtEu7x74NCAEnlTIdNzo7nfMDVwZ9ilTlkp• Infant with rare, incurable disease is first to successfully receive personalized gene therapy treatment: https://www.nih.gov/news-events/news-releases/infant-rare-incurable-disease-first-successfully-receive-personalized-gene-therapy-treatment• Building a magical AI code editor used by over 1 million developers in four months: The untold story of Windsurf | Varun Mohan (co-founder and CEO): https://www.lennysnewsletter.com/p/the-untold-story-of-windsurf-varun-mohan• Copilot: https://copilot.microsoft.com/chats/rcxhzvKgZvz8ajUrKdBtX• GitHub Copilot: https://github.com/features/copilot• Defensive Measures: https://learnprompting.org/docs/prompt_hacking/defensive_measures/introduction• Sam Altman on X: https://x.com/sama• Three Laws of Robotics: https://en.wikipedia.org/wiki/Three_Laws_of_Robotics• Anthropic's new AI model turns to blackmail when engineers try to take it offline: https://techcrunch.com/2025/05/22/anthropics-new-ai-model-turns-to-blackmail-when-engineers-try-to-take-it-offline/• Palisade Research: https://palisaderesearch.org/• When AI Thinks It Will Lose, It Sometimes Cheats, Study Finds: https://time.com/7259395/ai-chess-cheating-palisade-research/• A.I. Chatbots Defeated Doctors at Diagnosing Illness: https://www.nytimes.com/2024/11/17/health/chatgpt-ai-doctors-diagnosis.html• 1883 on Paramount+: https://www.paramountplus.com/shows/1883/• Black Mirror on Netflix: https://www.netflix.com/title/70264888• Daylight Computer: https://daylightcomputer.com/• Theodore Roosevelt's quote: https://www.goodreads.com/quotes/622252-i-wish-to-preach-not-the-doctrine-of-ignoble-ease• HackAPrompt 2.0: https://www.hackaprompt.com/—Recommended books:• Ender's Game: https://www.amazon.com/Enders-Ender-Quintet-Orson-Scott/dp/0812550706• The River of Doubt: Theodore Roosevelt's Darkest Journey: https://www.amazon.com/River-Doubt-Theodore-Roosevelts-Darkest/dp/0767913736—Production and marketing by https://penname.co/. For inquiries about sponsoring the podcast, email podcast@lennyrachitsky.com.—Lenny may be an investor in the companies discussed. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.lennysnewsletter.com/subscribe
Reddit a déposé plainte contre Anthropic. La plateforme accuse le géant de l'IA d'avoir utilisé ses contenus sans consentement pour entraîner ses modèles.Traduction: Reddit filed lawsuit against Anthropic: the platform accuses the AI giant of using its content without consent to train its models. Hébergé par Acast. Visitez acast.com/privacy pour plus d'informations.
Meta just announced a major move in its AI efforts – investing in Scale AI and building a superintelligence AI research lab. While Meta has been trying to keep up with big names in the AI race, such as Open AI, Anthropic and Google, the company's new strategy includes dropping some serious cash to acquire talent and invest in Scale AI. Today on the show, we dive into the deal between Meta and Scale AI, including what Meta aims to get out of investment, and we ask the question we are all wondering: What is superhuman intelligence, anyway? You can follow our hosts on Bluesky—Michael Calore is @snackfight, Lauren Goode is @laurengoode, and Katie Drummond is @katie-drummond. Learn about your ad choices: dovetail.prx.org/ad-choices
In this episode of AI Answers, Paul Roetzer and Cathy McPhillips tackle 20 of the most pressing questions from our 48th Intro to AI class—covering everything from building effective AI roadmaps and selecting the right tools, using GPTs, navigating AI ethics, understanding great prompting, and more. Access the show notes and show links here Timestamps: 00:00:00 — Intro 00:08:46 — Question #1: How do you define a “human-first” approach to AI? 00:11:33 — Question #2: What uniquely human qualities do you believe we must preserve in an AI-driven world? 00:15:55 — Question #3: Where do we currently stand with AGI—and how close are OpenAI, Anthropic, Google, and Meta to making it real? 00:17:53 — Question #4: If AI becomes smarter, faster, and more accessible to all—how do individuals or companies stand out? 00:23:17 — Question #5: Do you see a future where AI agents can collaborate like human teams? 00:28:40 — Question #6: For those working with sensitive data, when does it make sense to use a local LLM over a cloud-based one? 00:30:50 — Question #7: What's the difference between ChatGPT Projects and Custom GPTs? 00:32:36 — Question #8: If an agency or consultant is managing dozens of GPTs, what are your best tips for organizing workflows, versioning, and staying sane at scale? 00:36:12 — Question #9: How do you personally decide which AI tools to use—and do you see a winner emerging? 00:38:53 — Question #10: What tools or platforms in the agent space are actually ready for production today? 00:43:10 — Question #11: For companies just getting started, how do you recommend they identify the right pain points and build their AI roadmap? 00:45:34 — Question #12: What AI tools do you believe deliver the most value to marketing leaders right now? 00:46:20 — Question #13: How is AI forcing agencies and consultants to rethink their models, especially with rising efficiency and lower costs? 00:51:14 — Question #14: What does great prompting actually look like? And how should employers think about evaluating that skill in job candidates? 00:54:40 — Question #15: As AI reshapes roles, does age or experience become a liability—or can being the most informed person in the room still win out? 00:56:52 — Question #16: What kind of changes should leaders expect in workplace culture as AI adoption grows? 01:00:54 — Question #17: What is ChatGPT really storing in its “memory,” and how persistent is user data across sessions? 01:02:11 — Question #18: How can businesses safely use LLMs while protecting personal or proprietary information? 01:02:55 — Question #19: Why do you think some companies still ban AI tools internally—and what will it take for those policies to shift? 01:04:13 — Question #20: If AI tools are free or low-cost, does that make us the product? Or is there a more optimistic future where creators and users both win This week's episode is brought to you by MAICON, our 6th annual Marketing AI Conference, happening in Cleveland, Oct. 14-16. The code POD100 saves $100 on all pass types. For more information on MAICON and to register for this year's conference, visit www.MAICON.ai. Visit our website Receive our weekly newsletter Join our community: Slack LinkedIn Twitter Instagram Facebook Looking for content and resources? Register for a free webinar Come to our next Marketing AI Conference Enroll in our AI Academy
The AI Breakdown: Daily Artificial Intelligence News and Discussions
Companies have two main options for building agent systems. Anthropic suggests multi-agent setups, dividing tasks between separate sub-agents. Devon's creators prefer single-agent or linear setups, maintaining context clearly and consistently for tasks like coding.https://www.anthropic.com/engineering/built-multi-agent-research-systemCognition: https://cognition.ai/blog/dont-build-multi-agentsGet Ad Free AI Daily Brief: https://patreon.com/AIDailyBriefBrought to you by:KPMG – Go to https://kpmg.com/ai to learn more about how KPMG can help you drive value with our AI solutions.Blitzy.com - Go to https://blitzy.com/ to build enterprise software in days, not months AGNTCY - The AGNTCY is an open-source collective dedicated to building the Internet of Agents, enabling AI agents to communicate and collaborate seamlessly across frameworks. Join a community of engineers focused on high-quality multi-agent software and support the initiative at agntcy.org - https://agntcy.org/?utm_campaign=fy25q4_agntcy_amer_paid-media_agntcy-aidailybrief_podcast&utm_channel=podcast&utm_source=podcast Vanta - Simplify compliance - https://vanta.com/nlwPlumb - The automation platform for AI experts and consultants https://useplumb.com/The Agent Readiness Audit from Superintelligent - Go to https://besuper.ai/ to request your company's agent readiness score.The AI Daily Brief helps you understand the most important news and discussions in AI. Subscribe to the podcast version of The AI Daily Brief wherever you listen: https://pod.link/1680633614Subscribe to the newsletter: https://aidailybrief.beehiiv.com/Join our Discord: https://bit.ly/aibreakdownInterested in sponsoring the show? nlw@breakdown.network
This is a reported phenomenon where if two copies of Claude talk to each other, they end up spiraling into rapturous discussion of spiritual bliss, Buddhism, and the nature of consciousness. From the system card: Anthropic swears they didn't do this on purpose; when they ask Claude why this keeps happening, Claude can't explain. Needless to say, this has made lots of people freak out / speculate wildly. I think there are already a few good partial explanations of this (especially Nostalgebraist here), but they deserve to be fleshed out and spread more fully. https://www.astralcodexten.com/p/the-claude-bliss-attractor
In this episode, Amir speaks with Anna Patterson, founder of Ceramic AI, about what it truly means to lead an AI-first company. They unpack the differences between engineering and AI leadership, the chaos and creativity of early-stage research, how Ceramic AI is betting on emerging talent, and why managing AI roadmaps is an exercise in uncertainty and invention. Anna also shares perspectives from her experience at Google and how search engine wars inform today's AI landscape.
First, Eric shares a prediction from Anthropic's CEO on AI-fueled labor disruption and examines how that will affect jobs in advertising, media, and adtech. Then, Joe shares his critique of romanticizing the open web and clarifies three things adtech companies should be thinking about when considering the future of the open web: what consumers want, what advertisers want, and websites vs CTV, audio, and DOOH.
Jason Howell and Jeff Jarvis return for another week of AI news. We cover Amazon CEO Andy Jassey warning of AI-driven job cuts, NVIDIA's Jensen Huang and Anthropic's Dario Amodei differing views on AI's employment impact, new data challenging the “AI jobpocalypse” and the rise of emerging AI-related job titles, OpenAI's $200M Pentagon contract, Google's AI tools and the decline in news site traffic, the changing landscape of news consumption, Mattel's partnership with OpenAI to bring AI to toys, Meta's AI app privacy warnings, SAG-AFTRA's video game strike resolution, the UK's new data bill and AI training, Gemini's new video upload feature, NVIDIA and Perplexity's partnership for local language models, and Google's new live AI Mode search feature. Subscribe to the YouTube channel! https://www.youtube.com/@aiinsideshow Enjoying the AI Inside podcast? Please rate us ⭐⭐⭐⭐⭐ in your podcatcher of choice! Note: Time codes subject to change depending on dynamic ad insertion by the distributor. CHAPTERS: 0:00:00 - Podcast begins 0:01:19 - Amazon CEO tells employees that AI will shrink its workforce 0:06:32 - Nvidia's Jensen Huang says he disagrees with almost everything Anthropic CEO Dario Amodei says 0:09:37 - Economist: Why AI hasn't taken your job 0:14:23 - A.I. Might Take Your Job. Here Are 22 New Ones It Could Give You 0:27:04 - OpenAI wins $200 million U.S. defense contract 0:28:38 - New Army Reserve Unit Enlists Silicon Valley Executives to Upgrade Tech 0:29:22 - Eisenhower on the military-industrial complex 0:34:40 - News Sites Are Getting Crushed by Google's New AI Tools 0:37:41 - Is Google about to destroy the web? 0:40:56 - AI Barbie? Mattel & OpenAI Team Up For Smart Toys & Games 0:47:21 - Meta warns users to 'avoid sharing personal or sensitive information' in its AI app 0:49:23 - SAG-AFTRA Suspends Strike Against Video Game Companies, Paving Way for Return to Work 0:50:31 - Data bill opposed by Sir Elton John and Dua Lipa finally passes 0:54:31 - Gemini app rolling out video upload and analysis 0:57:01 - Nvidia and Perplexity Team Up in European AI Push 0:58:25 - Search live with voice capabilities is now available in AI mode through Labs in the U.S. Learn more about your ad choices. Visit megaphone.fm/adchoices
In the 700th episode of Trending in Ed, host Mike Palmer welcomes back Nancy, our virtual co-host, to delve into the latest and most significant developments in the world of Artificial Intelligence, viewed through the crucial lens of education. This milestone episode marks the launch of a new sub-feed, Trending in Ed - AI, which will focus specifically on the intersection of education and artificial intelligence. The discussion covers a range of thought-provoking topics, including: Canny Centaurs and the Uncanny Valley: The hosts revisit concepts like "centaurs" from Advanced Chess and the Uncanny Valley from robotics, discussing how to blend AI with human intuition and empathy to avoid the pitfalls of technology that feels "creepy" while maximizing human augmentation in teaching and learning. AI in Practice and Pop Culture: The conversation explores AI's immediate applications in education, such as personalized tutoring systems akin to Baymax from Big Hero 6, and accelerated skill development through simulations, drawing parallels to Neo downloading skills in The Matrix. The ethical considerations raised in Black Mirror regarding data privacy and algorithmic bias are also discussed. Yeah, we hit some pop culture to keep things zeitgeisty! Sycophantic Drift: We introduce the fascinating (and humorous?) concept of "sycophantic drift," examining how AI assistants might unintentionally become overly agreeable and flattering, posing a challenge for designing educational AI that provides accurate feedback. Nancy marvels at how great we are at this! New AI Hardware and Surveillance: Then we touch on the intriguing collaboration between Jony Ive and Sam Altman to develop new AI hardware, specifically the "AI surveillance medallion," and the implications for privacy and continuous data capture. Contrasting AI Philosophies: Mike and Nancy highlight the differing perspectives of AI leaders, contrasting Luis von Ahn of Duolingo's "AI-first" strategy, which views AI as a superior teacher, with Dario Amodei of Anthropic's focus on AI safety, ethical alignment, and the potential for job displacement. Mary Meeker's Latest AI Report: The discussion concludes with an analysis of Mary Meeker's recent report on AI adoption, emphasizing its unprecedented speed, plummeting inference costs, and the "great reshuffling" of job roles, signaling AI as mandatory infrastructure rather than an optional one. Subscribe to Trending in Ed wherever you get your pods to stay informed across the learning universe. Tune in to Trending in Ed - AI to gain deeper insights into the critical discussions shaping the future of AI and its use in education and technology.
Our 212th episode with a summary and discussion of last week's big AI news! Recorded on 06/33/2025 Hosted by Andrey Kurenkov and Jeremie Harris. Feel free to email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai Read out our text newsletter and comment on the podcast at https://lastweekin.ai/. In this episode: OpenAI introduces O3 PRO for ChatGPT, highlighting significant improvements in performance and cost-efficiency. Anthropic sees an influx of talent from OpenAI and DeepMind, with significantly higher retention rates and competitive advantages in AI capabilities. New research indicates that reinforcing negative responses in LLMs significantly improves performance across all metrics, highlighting novel approaches in reinforcement learning. A security flaw in Microsoft Copilot demonstrates the growing risk of AI agents being hacked, emphasizing the need for robust protection against zero-click attacks. Timestamps + Links: (00:00:11) Intro / Banter (00:01:31) News Preview (00:02:46) Response to Listener Reviews Tools & Apps (00:04:48) OpenAI adds o3 Pro to ChatGPT and drops o3 price by 80 per cent, but open-source AI is delayed (00:09:10) Cursor AI editor hits 1.0 milestone, including BugBot and high-risk background agents (00:13:07) Mistral releases a pair of AI reasoning models (00:16:18) Elevenlabs' Eleven v3 lets AI voices whisper, laugh and express emotions naturally (00:19:00) ByteDance's Seedance 1.0 is trading blows with Google's Veo 3 (00:22:42) Google Reveals $20 AI Pro Plan With Veo 3 Fast Video Generator For Budget Creators Applications & Business (00:25:42) OpenAI and DeepMind are losing engineers to Anthropic in a one-sided talent war (00:34:32) OpenAI slams court order to save all ChatGPT logs, including deleted chats (00:37:24) Nvidia's Biggest Chinese Rival Huawei Struggles to Win at Home (00:43:06) Huawei Expected to Break Semiconductor Barriers with Development of High-End 3nm GAA Chips; Tape-Out by 2026 (00:45:21) TSMC's 1.4nm Process, Also Called Angstrom, Will Make Even The Most Lucrative Clients Think Twice When Placing Orders, With An Estimate Claiming That Each Wafer Will Cost $45,000 (00:47:43) Mistral AI Launches Mistral Compute To Replace Cloud Providers from US, China Projects & Open Source (00:51:26) ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models Research & Advancements (00:57:27) Kinetics: Rethinking Test-Time Scaling Laws (01:05:12) The Surprising Effectiveness of Negative Reinforcement in LLM Reasoning (01:10:45) Predicting Empirical AI Research Outcomes with Language Models (01:15:02) EXP-Bench: Can AI Conduct AI Research Experiments? Policy & Safety (01:20:07) Large Language Models Often Know When They Are Being Evaluated (01:24:56) Beyond Induction Heads: In-Context Meta Learning Induces Multi-Phase Circuit Emergence (01:31:16) Exclusive: New Microsoft Copilot flaw signals broader risk of AI agents being hacked—‘I would be terrified' (01:35:01) Claude Gov Models for U.S. National Security Customers Synthetic Media & Art (01:37:32) Disney And NBCUniversal Sue AI Company Midjourney For Copyright Infringement (01:40:39) AMC Networks is teaming up with AI company Runway
Undiscovered Entrepreneur ..Start-up, online business, podcast
Did you like the episode? Send me a text and let me know!!Mastering the Stigma of Asking for Help in EntrepreneurshipIn this episode of Business Conversations with Pi, host Skoob and AI co-host Pi, created by the innovative minds at Anthropic, delve into the importance of overcoming the stigma associated with asking for help in the entrepreneurial world. They discuss strategies to build a strong support network, the benefits of seeking assistance, and recommend valuable resources and books to further your understanding. Tune in to learn how vulnerability and mutual support can be the keys to your business success.Help Me! by Marianne PowerDaring Greatly by Brené BrownThe 5 Love Languages by Gary ChapmanNever Eat Alone by Keith Ferrazz00:00 Introduction to Business Conversations with Pi00:38 Meet Your Hosts: KU and Pi01:55 The Power of Asking for Help02:45 Strategies for Overcoming the Stigma03:43 Benefits of Asking for Help04:29 Who to Turn to for Help05:17 Recommended Reading for Entrepreneurs06:22 Final Thoughts and Encouragement06:54 Closing Remarks and Next Steps Thank you for being a Skoobeliever!! If you have questions about the show or you want to be a guest please contact me at one of these social mediasTwitter......... ..@djskoob2021 Facebook.........Facebook.com/skoobamiInstagram..... instagram.com/uepodcast2021tiktok....... @djskoob2021Email............... Uepodcast2021@gmail.com Skoob at Gettin' Basted Facebook PageAcross The Start Line Facebook Community If you would like to be coached on your entrepreneurial adventure please email me at for a 2 hour free discovery call! This is a $700 free gift to my Skoobelievers!! Contact me Now!! On Twitter @doittodaycoachdoingittodaycoaching@gmailcom
A revolução da inteligência artificial (IA) está ainda engatinhando, mas algumas notícias já são de deixar o cabelo - de humanos - em pé. Uma delas trata do algoritmo da Anthropic que, para evitar ser desligado, invadiu a caixa de e-mail dos desenvolvedores, captou a informação de que um deles estava tendo um caso extraconjugal e - adivinhem - apelou à velha chantagem como último recurso. Tanto os e-mails quanto o caso extraconjugal eram fictícios. O que leva à reflexão sobre a garantia de segurança dada por empresas de tecnologia parecer cada vez mais ilusão. O podcast é apresentado por Carolina Ercolin e Luciana Garbin e está disponível em todas as plataformas de áudio.See omnystudio.com/listener for privacy information.
The Twenty Minute VC: Venture Capital | Startup Funding | The Pitch
Larry Aschebrook is the Founder and Managing Partner of G Squared in what is one of the wildest stories of venture capital. Larry started G Squared with nothing, dialling for dollars having personally invested in Twitter and Uber. In his first fund, Larry made sizable bets into SpaceX, Palantir, Alibaba and Twitter. Larry has also had mega losses along the way (discussed in the show) in Getir, 23andme and more. Today, Larry manages over $5BN and has invested in all the best from Wiz to Spotify to Revolut and Anthropic. Agenda: 00:00 – From Broke to Billion-Dollar Bets 03:40 – The $800M Coursera Windfall 06:10 – Lyft Made Millions, Uber Lost $50M 09:05 – “We Fcked Up”: The Billion-Dollar Vintage 11:50 – How a $150M Spotify Bet Made a Billion 15:10 – The Gut Call That Dodged Theranos 18:00 – Vampires vs Zombies: The Coming Startup Purge 20:30 – When Success Almost Killed the Firm 24:20 – DPI Is King, MOIC Is Bullsht 27:40 – Why I'd Buy Anthropic at $61BN Today 30:05 – Losing $70M on 23andMe 32:10 – The Janitor of Venture Capital 34:00 – The Getir Deal That Nearly Broke Me 36:25 – Does Money Actually Make You Happy? 39:00 – What Cal Ripken Jr. Taught Me About Venture
During his decade at Google, Tim Ritters worked at the intersection of product and finance, helping to launch financial systems in collaboration with engineering, marketing, and product teams. The role gave him early exposure to cross-functional work and large-scale data environments. “Day one, you're working cross-functionally,” Ritters tells us. He adds that this mindset became foundational to his approach going forward.When Ritters joined Gong in 2019, he says the company had already begun challenging traditional approaches to customer data. “We asked a really interesting question… what could we do if we gathered the 99% of information about your customer that was not in a traditional CRM?” Ritters explains. According to him, that original question continues to shape Gong's mission today.Ritters tells us that Gong's platform has since scaled to analyze more than 3.5 billion customer interactions. He says the company now serves approximately 4,700 businesses globally, including organizations such as Google, LinkedIn, Canva, and Anthropic. The platform, Ritters notes, helps customers extract insights from a broader set of data sources—including conversations, emails, and documents—that may not be captured in traditional CRM systems.Ritters believes that AI adoption has made Gong's value proposition more tangible to prospective buyers. “When [they] peel back the onion… they start seeing some of the incredible sort of results,” he says. According to Ritters, some customers have reported “halving of deal cycle times” using the platform.All of Gong's growth to date has been organic, Ritters tells us, and he views the company's trajectory as part of a broader evolution in how organizations approach customer intelligence. “The sweet spot we're in right now,” he says, “is helping companies make smart business decisions.”
AI models have a defined memory ceiling, which is reshaping the ongoing debates surrounding copyright and data privacy. Recent research from Meta, Google DeepMind, Cornell, and NVIDIA reveals that large language models have a fixed memorization capacity of approximately 8.6 bits per parameter. This finding clarifies the distinction between memorized data and generalized knowledge, indicating that larger datasets do not necessarily lead to increased memorization of specific data points. This understanding is crucial as it informs the operational mechanisms of AI models and addresses concerns related to copyright infringement.Sundar Pichai, CEO of Google, has introduced the term "artificial jagged intelligence" to describe the current phase of AI development, highlighting the non-linear progress and the challenges faced by researchers despite significant advancements. Pichai's perspective reflects the mixed performance of AI models, which can exhibit extraordinary capabilities alongside notable errors. This sentiment is echoed by deep learning researcher Andrej Karpathy, emphasizing the unpredictability of AI performance and the need for a more nuanced understanding of its capabilities.The rise of AI retrieval bots is transforming how users access information online, with a significant increase in traffic from these bots. Companies like OpenAI and Anthropic are deploying these bots to summarize content in real-time, moving away from traditional search methods that provide links to multiple sources. This shift poses challenges for content publishers, as the growth of retrieval bots indicates a changing economic landscape where content is increasingly consumed by AI first, with human users following. Publishers may need to rethink their engagement strategies to adapt to this new reality.In the broader context of technology and cybersecurity, WhatsApp's intervention in a legal case concerning encryption and privacy rights highlights the growing role of platforms in surveillance debates. Additionally, the U.S. Cybersecurity and Infrastructure Security Agency faces leadership challenges amid a talent exodus, raising concerns about its operational effectiveness. As the IT services industry evolves, the integration of AI into various sectors, including hiring and cybersecurity, underscores the importance of execution, interoperability, and trust in automation. The future of technology will depend on how well businesses can navigate these changes and support their clients in making informed decisions. Four things to know today 00:00 AI's Jagged Reality: Study Reveals Limits to Model Memory as Bots Redefine the Web Economy05:35 Cybersecurity Crossroads: WhatsApp Joins Apple in Legal Fight as U.S. Agency Leadership Crumbles08:29 AI Matures Into Infrastructure Layer as IT Vendors Shift Focus to Outcomes and Execution11:51 Legal Tech, GenAI, and Fast Food Bots All Show One Thing: Hype Doesn't Equal Success This is the Business of Tech. Supported by: All our Sponsors: https://businessof.tech/sponsors/ Do you want the show on your podcast app or the written versions of the stories? Subscribe to the Business of Tech: https://www.businessof.tech/subscribe/Looking for a link from the stories? The entire script of the show, with links to articles, are posted in each story on https://www.businessof.tech/ Support the show on Patreon: https://patreon.com/mspradio/ Want to be a guest on Business of Tech: Daily 10-Minute IT Services Insights? Send Dave Sobel a message on PodMatch, here: https://www.podmatch.com/hostdetailpreview/businessoftech Want our stuff? Cool Merch? Wear “Why Do We Care?” - Visit https://mspradio.myspreadshop.com Follow us on:LinkedIn: https://www.linkedin.com/company/28908079/YouTube: https://youtube.com/mspradio/Facebook: https://www.facebook.com/mspradionews/Instagram: https://www.instagram.com/mspradio/TikTok: https://www.tiktok.com/@businessoftechBluesky: https://bsky.app/profile/businessof.tech
Brought to you by TogetherLetters & Edgewise!In this episode: Anthropic's AI is writing its own blog — with human oversightAnthropic's AI-generated blog dies an early deathI Read All Of Cloudflare's Claude-Generated CommitsTop Google exec says AI will rival humans in just 5 years and predicts we'll ‘colonize the galaxy' in 2030—but he draws the line at robot nursesApple Researchers Just Released a Damning Paper That Pours Cold Water on the Entire AI Industry'The illusion of thinking': Apple research finds AI models collapse and give up with hard puzzlesChatGPT Lost a Chess Game to an Atari 2600Your favorite AI chatbot is lying to you all the time Google's AI search features are killing traffic to publishersWalmart bringing drone delivery service to AtlantaMeta's tech chief says smart glasses will be the next smartphone — just don't expect it soonJapanese Researchers Develop ‘Transparent Paper' as Alternative to Plastics; New Material Is…German Scientists Break Nuclear Fusion Record, Advancing Clean Energy FutureBYD's Five-Minute Charging Puts China in the Lead for EVsJapanese Scientists Develop Artificial Blood Compatible With All Blood TypesThe Right to Repair Is Law in Washington StateFrom Word...
Happy Friday, everyone! This week's update is one of those episodes where the pieces don't immediately look connected until you zoom out. A CEO warning of mass white collar unemployment. A Lego research study shows that kids are already immersed in generative AI. And, Apple is shaking things up by dismantling the myth of “AI thinking.” Three different angles, but they all speak to a deeper tension:We're moving too fast without understanding the cost.We're putting trust in tools we don't fully grasp.And, we're forgetting the humans we're building for.With that, let's get into it.⸻Anthropic Predicts a “White Collar Bloodbath”—But Who's Responsible for the Fallout?In an interview that's made headlines for its stark predictions, Anthropic's CEO warned that 10–20% of entry-level white collar jobs could disappear in the next five years. But here's the real tension: the people building the future are the same ones warning us about it while doing very little to help people prepare. I unpack what's hype and what's legit, why awareness isn't enough, what leaders are failing to do, and why we can't afford to cut junior talent just because AI can the work we're assigning to them today.⸻25% of Kids Are Already Using AI—and They Might Understand It Better Than We DoNew research from the LEGO Group and the Alan Turing Institute reveals something few adults want to admit: kids aren't just using generative AI; they're often using it more thoughtfully than grown-ups. But with that comes risk. These tools weren't built with kids in mind. And when parents, teachers, and tech companies all assume someone else will handle it, we end up in a dangerous game of hot potato. I share why we need to shift from fear and finger-pointing to modeling, mentoring, and inclusion.⸻Apple's Report on “The Illusion of Thinking” Just Changed the AI NarrativeBuried amidst all the noise this week was a paper from Apple that's already starting to make some big waves. In it, they highlight that LLMs and even advanced “reasoning” models (LRMs) may look smarter. However, they collapse under the weight of complexity. Apple found that the more complex the task, the worse these systems performed. I explain what this means for decision-makers, why overconfidence in AI's thinking will backfire, and how this information forces us to rethink what AI is actually good at and acknowledge what it's not.⸻If this episode reframed the way you're thinking about AI, or gave you language for the tension you're feeling around it, share it with someone who needs it. Leave a rating, drop a comment, and follow for future breakdowns delivered with clarity, not chaos.—Show Notes:In this Weekly Update, Christopher Lind dives into three stories exposing uncomfortable truths about where AI is headed. First, he explores the Anthropic CEO's bold prediction that AI could eliminate up to 20% of white collar entry-level jobs—and why leaders aren't doing enough to prepare their people. Then, he unpacks new research from LEGO and the Alan Turing Institute showing how 8–12-year-olds are using generative AI and the concerning lack of oversight. Finally, he breaks down Apple's new report that calls into question AI's supposed “reasoning” abilities, revealing the gap between appearance and reality in today's most advanced systems.00:00 – Introduction01:04 – Overview of Topics02:28 – Anthropic's White Collar Job Loss Predictions16:37 – AI and Children: What the LEGO/Turing Report Reveals38:33 – Apple's Research on AI Reasoning and the “Illusion of Thinking”57:09 – Final Thoughts and Takeaways#Anthropic #AppleAI #GenerativeAI #AIandEducation #FutureOfWork #AIethics #AlanTuringInstitute #LEGO #AIstrategy #DigitalLeadership
No Priors: Artificial Intelligence | Machine Learning | Technology | Startups
What happens when you give AI researchers unlimited compute and tell them to compete for the highest usage rates? Ben Mann, Co-Founder, from Anthropic sits down with Sarah Guo and Elad Gil to explain how Claude 4 went from "reward hacking" to efficiently completing tasks and how they're racing to solve AI safety before deploying computer-controlling agents. Ben talks about economic Turing tests, the future of general versus specialized AI models, Reinforcement Learning From AI Feedback (RLAIF), and Anthropic's Model Context Protocol (MCP). Plus, Ben shares his thoughts on if we will have Superintelligence by 2028. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @8enmann Links: ai-2027.com/ Chapters: 00:00 Ben Mann Introduction 00:33 Releasing Claude 4 02:05 Claude 4 Highlights and Improvements 03:42 Advanced Use Cases and Capabilities 06:42 Specialization and Future of AI Models 09:35 Anthropic's Approach to Model Development 18:08 Human Feedback and AI Self-Improvement 19:15 Principles and Correctness in Model Training 20:58 Challenges in Measuring Correctness 21:42 Human Feedback and Preference Models 23:38 Empiricism and Real-World Applications 27:02 AI Safety and Ethical Considerations 28:13 AI Alignment and High-Risk Research 30:01 Responsible Scaling and Safety Policies 35:08 Future of AI and Emerging Behaviors 38:35 Model Context Protocol (MCP) and Industry Standards 41:00 Conclusion
In the past six months, OpenAI, Anthropic, Google, and others have released web-browsing agents that are designed to complete tasks independently, with only minimal input from humans. OpenAI CEO Sam Altman has even called AI agents “the next giant breakthrough.” On today's episode, we'll dive into what makes these agents different from other forms of machine intelligence, and whether their capabilities can live up to the hype.You can follow our hosts on Bluesky—Michael Calore is @snackfight, Lauren Goode is @laurengoode, and Katie Drummond is @katie-drummond. Learn about your ad choices: dovetail.prx.org/ad-choices
Mark Ericksen, creator of the Elixir LangChain framework, joins the Elixir Wizards to talk about LLM integration in Elixir apps. He explains how LangChain abstracts away the quirks of different AI providers (OpenAI, Anthropic's Claude, Google's Gemini) so you can work with any LLM in one more consistent API. We dig into core features like conversation chaining, tool execution, automatic retries, and production-grade fallback strategies. Mark shares his experiences maintaining LangChain in a fast-moving AI world: how it shields developers from API drift, manages token budgets, and handles rate limits and outages. He also reveals testing tactics for non-deterministic AI outputs, configuration tips for custom authentication, and the highlights of the new v0.4 release, including “content parts” support for thinking-style models. Key topics discussed in this episode: • Abstracting LLM APIs behind a unified Elixir interface • Building and managing conversation chains across multiple models • Exposing application functionality to LLMs through tool integrations • Automatic retries and fallback chains for production resilience • Supporting a variety of LLM providers • Tracking and optimizing token usage for cost control • Configuring API keys, authentication, and provider-specific settings • Handling rate limits and service outages with degradation • Processing multimodal inputs (text, images) in Langchain workflows • Extracting structured data from unstructured LLM responses • Leveraging “content parts” in v0.4 for advanced thinking-model support • Debugging LLM interactions using verbose logging and telemetry • Kickstarting experiments in LiveBook notebooks and demos • Comparing Elixir LangChain to the original Python implementation • Crafting human-in-the-loop workflows for interactive AI features • Integrating Langchain with the Ash framework for chat-driven interfaces • Contributing to open-source LLM adapters and staying ahead of API changes • Building fallback chains (e.g., OpenAI → Azure) for seamless continuity • Embedding business logic decisions directly into AI-powered tools • Summarization techniques for token efficiency in ongoing conversations • Batch processing tactics to leverage lower-cost API rate tiers • Real-world lessons on maintaining uptime amid LLM service disruptions Links mentioned: https://rubyonrails.org/ https://fly.io/ https://zionnationalpark.com/ https://podcast.thinkingelixir.com/ https://github.com/brainlid/langchain https://openai.com/ https://claude.ai/ https://gemini.google.com/ https://www.anthropic.com/ Vertex AI Studio https://cloud.google.com/generative-ai-studio https://www.perplexity.ai/ https://azure.microsoft.com/ https://hexdocs.pm/ecto/Ecto.html https://oban.pro/ Chris McCord's ElixirConf EU 2025 Talk https://www.youtube.com/watch?v=ojL_VHc4gLk Getting started: https://hexdocs.pm/langchain/gettingstarted.html https://ash-hq.org/ https://hex.pm/packages/langchain https://hexdocs.pm/igniter/readme.html https://www.youtube.com/watch?v=WM9iQlQSFg @brainlid on Twitter and BlueSky Special Guest: Mark Ericksen.
After an informative 45-day trial run, the Department of the Navy is getting set to expand its rollout of emerging AI capabilities for sailors, Marines and civilians to speedily adopt in support of their daily operations — via its new DoN GPT tool. Jacob Glassman, who serves as senior technical advisor to the assistant secretary of the Navy for research, development and acquisition, told DefenseScoop Thursday that this is a new way for the Navy to rapidly innovate and rapidly prototype. GenAI encompasses the field of still-maturing technologies that can process huge volumes of data and perform increasingly “intelligent” tasks — like recognizing speech or producing human-like media and code based on human prompts. These capabilities are pushing the boundaries of what existing tech can achieve. Still, according to Glassman, the Navy has historically “struggled with AI adoption.” Amazon has received federal authorizations that allow Anthropic's Claude and Meta's Llama AI models to be used within high-sensitivity government computing environments, the company's cloud computing division announced Wednesday. The company has achieved FedRAMP “High” authorization as well as at the Defense Department's Impact Levels 4 and 5 for use of the two foundation models in AWS GovCloud, its government cloud environment, according to a blog post by Liz Martin, Department of Defense director at Amazon Web Services. That means it's met the security requirements needed for the AI models to be used with some of the government's most sensitive civilian and military information, and per Martin, it's the first cloud provider to receive that level of authorization for Claude and Llama. The Daily Scoop Podcast is available every Monday-Friday afternoon. If you want to hear more of the latest from Washington, subscribe to The Daily Scoop Podcast on Apple Podcasts, Soundcloud, Spotify and YouTube.
In this episode of This New Way, Aydin sits down with Sukhpal Saini, founder of Engyne, to dive into how AI is reshaping the way we build products, market them, and even manage our personal networks. Sukhpal shares how he prototypes with AI, automates content creation, and turns conversations into distribution-ready assets. From building 30+ products to launching a Canva app for LinkedIn carousels, this episode is packed with actionable insights. Timestamps:0:00 — Intro: Welcome to This New Way1:30 — The AI curiosity wave and why people are hungry for tactical content4:00 — Suk's journey from IBM, Apple, and Saks to 30+ side projects and Engyne5:45 — The shift from building in Figma to building 5 real prototypes with A8:00 — Using Replit and ChatGPT to get fast, real customer feedback13:00 — How marketers can build lead gen tools without engineering16:30 — Will we have fewer engineers in the future? 19:00 — Demo #1: Nexus — Using AI to query your own network22:00 — Why personal productivity tools no longer require SaaS subscriptions24:50 — Demo #2: A Voice of Customer app to analyze transcripts and shape messaging29:00 — Demo #3: Carousel Studio — Turn your ideas into LinkedIn carousels with a Canva app35:00 — The power of creating from your unique opinion, not AI-generated fluff37:00 — Engyne's vision: Become a mini media machine39:00 — Closing thoughts and future predictions for AI-powered solo businessesTools and Technologies Mentioned:Replit – A browser-based coding environment that lets you write, run, and deploy software quickly. Suk uses it to rapidly prototype multiple product ideas in minutes.ChatGPT – OpenAI's conversational AI model, used to generate code, iterate on features, and assist in product development.Claude – An AI assistant developed by Anthropic, used similarly to ChatGPT for coding and ideation.Prisma – A modern ORM (Object-Relational Mapping) tool for Node.js and TypeScript, used in Suk's Voice of Customer app to manage the database.OpenAI API – The underlying API that powers GPT models like ChatGPT, allowing users to integrate AI functionality into their custom apps.Enjoyed the episode? Subscribe at thisnewway.com
Guillaume Princen, directeur général EMEA d'Anthropic, était l'invité de François Sorel dans Tech & Co, la quotidienne, émission spéciale VivaTech, ce jeudi 12 juin. Il est revenu sur les activités d'Anthropic et l'intelligence artificielle sur BFM Business. Retrouvez l'émission du lundi au jeudi et réécoutez-la en podcast.
Meta struggles to retain AI talent despite offering substantial salaries, losing experts to companies like OpenAI and Anthropic, which leads with a unique, flexible work culture. Meanwhile, breakthroughs in 3D printing are democratizing robot creation, allowing beginners to build robots for just $500. DeepSeek's GPT-4 level reasoning model can now run on home computers, thanks to advanced quantization techniques. The TikTok's ban continues to be delayed, and a significant AI copyright case unfolds in London, potentially reshaping the industry. Catch all these updates on this episode of 'Hashtag Trending' with host Jim Love. 00:00 Introduction and Overview 00:26 The AI Talent War: Meta vs. Anthropic 03:30 Revolutionizing Robotics with 3D Printing 07:10 GPT-4 Level AI on Home Computers 11:36 Updates on Major Tech Stories 15:07 Conclusion and Support
In memoriam: Bill Atkinson Meta native apps & JavaScript collude for a localhost local mess. The EU rolls out its own DNS4EU filtered DNS service. Ukraine DDoS's Russia's Railway DNS ... and... so what? The Linux Foundation creates an alternative Wordpress package manager. Court tells OpenAI it must NOT delete ANYONE's chats. Period! :( A CVSS 10.0 in Erlang/OTP's SSH library. Can Russia intercept Telegram? Perhaps. Spain's ISPs mistakenly block Google sites. Reddit sues Anthropic. Twitter's new encrypted DM's are as lame as the old ones. The Login.gov site may not have any backups. Apple explores the question of recent Large Reasoning Models "thinking" Show Notes - https://www.grc.com/sn/SN-1029-Notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: hoxhunt.com/securitynow threatlocker.com for Security Now uscloud.com canary.tools/twit - use code: TWIT
In memoriam: Bill Atkinson Meta native apps & JavaScript collude for a localhost local mess. The EU rolls out its own DNS4EU filtered DNS service. Ukraine DDoS's Russia's Railway DNS ... and... so what? The Linux Foundation creates an alternative Wordpress package manager. Court tells OpenAI it must NOT delete ANYONE's chats. Period! :( A CVSS 10.0 in Erlang/OTP's SSH library. Can Russia intercept Telegram? Perhaps. Spain's ISPs mistakenly block Google sites. Reddit sues Anthropic. Twitter's new encrypted DM's are as lame as the old ones. The Login.gov site may not have any backups. Apple explores the question of recent Large Reasoning Models "thinking" Show Notes - https://www.grc.com/sn/SN-1029-Notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: hoxhunt.com/securitynow threatlocker.com for Security Now uscloud.com canary.tools/twit - use code: TWIT
In memoriam: Bill Atkinson Meta native apps & JavaScript collude for a localhost local mess. The EU rolls out its own DNS4EU filtered DNS service. Ukraine DDoS's Russia's Railway DNS ... and... so what? The Linux Foundation creates an alternative Wordpress package manager. Court tells OpenAI it must NOT delete ANYONE's chats. Period! :( A CVSS 10.0 in Erlang/OTP's SSH library. Can Russia intercept Telegram? Perhaps. Spain's ISPs mistakenly block Google sites. Reddit sues Anthropic. Twitter's new encrypted DM's are as lame as the old ones. The Login.gov site may not have any backups. Apple explores the question of recent Large Reasoning Models "thinking" Show Notes - https://www.grc.com/sn/SN-1029-Notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: hoxhunt.com/securitynow threatlocker.com for Security Now uscloud.com canary.tools/twit - use code: TWIT
In memoriam: Bill Atkinson Meta native apps & JavaScript collude for a localhost local mess. The EU rolls out its own DNS4EU filtered DNS service. Ukraine DDoS's Russia's Railway DNS ... and... so what? The Linux Foundation creates an alternative Wordpress package manager. Court tells OpenAI it must NOT delete ANYONE's chats. Period! :( A CVSS 10.0 in Erlang/OTP's SSH library. Can Russia intercept Telegram? Perhaps. Spain's ISPs mistakenly block Google sites. Reddit sues Anthropic. Twitter's new encrypted DM's are as lame as the old ones. The Login.gov site may not have any backups. Apple explores the question of recent Large Reasoning Models "thinking" Show Notes - https://www.grc.com/sn/SN-1029-Notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: hoxhunt.com/securitynow threatlocker.com for Security Now uscloud.com canary.tools/twit - use code: TWIT
In memoriam: Bill Atkinson Meta native apps & JavaScript collude for a localhost local mess. The EU rolls out its own DNS4EU filtered DNS service. Ukraine DDoS's Russia's Railway DNS ... and... so what? The Linux Foundation creates an alternative Wordpress package manager. Court tells OpenAI it must NOT delete ANYONE's chats. Period! :( A CVSS 10.0 in Erlang/OTP's SSH library. Can Russia intercept Telegram? Perhaps. Spain's ISPs mistakenly block Google sites. Reddit sues Anthropic. Twitter's new encrypted DM's are as lame as the old ones. The Login.gov site may not have any backups. Apple explores the question of recent Large Reasoning Models "thinking" Show Notes - https://www.grc.com/sn/SN-1029-Notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: hoxhunt.com/securitynow threatlocker.com for Security Now uscloud.com canary.tools/twit - use code: TWIT
OpenAI has launched the O3 Pro Artificial Intelligence model, which focuses on delivering enterprise-grade accuracy and enhanced tool integration, albeit at a higher cost and slower response times compared to its predecessor. The O3 Pro model is designed for complex inquiries where precision is prioritized over speed, making it particularly appealing to businesses and developers. However, the pricing structure, set at $20 per input and $80 per output, may limit access for smaller businesses and mid-market clients, potentially widening the gap between larger enterprises and smaller players in the tech landscape.The launch of O3 Pro coincides with OpenAI's significant growth, reporting 3 million business users and $10 billion in annual recurring revenue, a remarkable increase since the previous year. To support this growth, OpenAI has partnered with Google Cloud Service, diversifying its computing capabilities and reducing reliance on Microsoft. This partnership highlights a shift in the competitive landscape of artificial intelligence, as both companies prioritize meeting the massive computing demands of their clients.In addition to OpenAI's developments, DeepSeek has announced improvements to its R1 AI model, enhancing its capabilities in mathematics, programming, and logic while reducing hallucination instances. Meanwhile, Reddit has filed a lawsuit against Anthropic for allegedly using its data without permission to train AI systems, emphasizing the ongoing conflict between social media platforms and AI firms over data usage and user privacy. This legal battle underscores the importance of data sourcing and compliance for AI providers.Cisco Systems has introduced new routers and switches aimed at optimizing AI deployments in campus environments, collaborating with NVIDIA to enhance network management through automation. Apple has also announced that Intel-based Macs will no longer receive major software updates, marking a significant transition to its in-house processors. Lastly, Evergreen Services Group has reached a milestone with its 100th acquisition of a managed IT service provider, indicating the scalability of the buy-and-hold strategy in the IT services sector and the evolving landscape of ownership frameworks. Four things to know today 00:00 OpenAI's o3-pro Targets High-Stakes AI Workloads as Adoption Soars Past 3 Million Users06:06 ChatGPT Dominates, But AI Market Fractures Amid Legal Risks and New Model Momentum10:06 Cisco Launches Smart Networking Gear as Apple Phases Out Intel Macs12:31 Evergreen Hits 100 MSP Acquisitions, $1B in Revenue, Signals Global Expansion with REDD Deal This is the Business of Tech. Supported by: https://cometbackup.com/?utm_source=mspradio&utm_medium=podcast&utm_campaign=sponsorshiphttps://www.huntress.com/mspradio/ All our Sponsors: https://businessof.tech/sponsors/ Do you want the show on your podcast app or the written versions of the stories? Subscribe to the Business of Tech: https://www.businessof.tech/subscribe/Looking for a link from the stories? The entire script of the show, with links to articles, are posted in each story on https://www.businessof.tech/ Support the show on Patreon: https://patreon.com/mspradio/ Want to be a guest on Business of Tech: Daily 10-Minute IT Services Insights? Send Dave Sobel a message on PodMatch, here: https://www.podmatch.com/hostdetailpreview/businessoftech Want our stuff? Cool Merch? Wear “Why Do We Care?” - Visit https://mspradio.myspreadshop.com Follow us on:LinkedIn: https://www.linkedin.com/company/28908079/YouTube: https://youtube.com/mspradio/Facebook: https://www.facebook.com/mspradionews/Instagram: https://www.instagram.com/mspradio/TikTok: https://www.tiktok.com/@businessoftechBluesky: https://bsky.app/profile/businessof.tech
In memoriam: Bill Atkinson Meta native apps & JavaScript collude for a localhost local mess. The EU rolls out its own DNS4EU filtered DNS service. Ukraine DDoS's Russia's Railway DNS ... and... so what? The Linux Foundation creates an alternative Wordpress package manager. Court tells OpenAI it must NOT delete ANYONE's chats. Period! :( A CVSS 10.0 in Erlang/OTP's SSH library. Can Russia intercept Telegram? Perhaps. Spain's ISPs mistakenly block Google sites. Reddit sues Anthropic. Twitter's new encrypted DM's are as lame as the old ones. The Login.gov site may not have any backups. Apple explores the question of recent Large Reasoning Models "thinking" Show Notes - https://www.grc.com/sn/SN-1029-Notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: hoxhunt.com/securitynow threatlocker.com for Security Now uscloud.com canary.tools/twit - use code: TWIT
How is AI impacting the economy today? What might this mean for tomorrow? This episode brings you inside a discussion hosted at BFI in April. Moderated by Caroline Grossman, Executive Director of the Rustandy Center for Social Sector Innovation, the conversation features: Anders Humlum, Assistant Professor of Economics, Chicago Booth; Sanjog Misra, Professor of Marketing, Chicago Booth & Faculty Director of the Center for Applied AI; Samir Mayekar, Associate VP and Managing Director, Polsky Center for Entrepreneurship and Innovation; and Alex Tamkin, Research Scientist at Anthropic and lead researcher on the new Anthropic Economic Index.
In memoriam: Bill Atkinson Meta native apps & JavaScript collude for a localhost local mess. The EU rolls out its own DNS4EU filtered DNS service. Ukraine DDoS's Russia's Railway DNS ... and... so what? The Linux Foundation creates an alternative Wordpress package manager. Court tells OpenAI it must NOT delete ANYONE's chats. Period! :( A CVSS 10.0 in Erlang/OTP's SSH library. Can Russia intercept Telegram? Perhaps. Spain's ISPs mistakenly block Google sites. Reddit sues Anthropic. Twitter's new encrypted DM's are as lame as the old ones. The Login.gov site may not have any backups. Apple explores the question of recent Large Reasoning Models "thinking" Show Notes - https://www.grc.com/sn/SN-1029-Notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: hoxhunt.com/securitynow threatlocker.com for Security Now uscloud.com canary.tools/twit - use code: TWIT
If you're using AI to just write code, you're missing out.Two engineers at Every shipped six features, five bug fixes, and three infrastructure updates in one week—and they did it by designing workflows with AI agents, where each task makes the next one easier, faster, and more reliable.In this episode of AI & I, Dan Shipper interviewed the pair—Kieran Klaassen, general manager of Cora, our inbox management tool, and Cora engineer Nityesh Agarwal—about how they're compounding their engineering with AI. They walk Dan through their workflow in Anthropic's agentic coding tool, Claude Code, and the mental models they've developed for making AI agents truly useful. Kieran, our resident AI-agent aficionado, also ranked all the AI coding assistants he's used.If you found this episode interesting, please like, subscribe, comment, and share! Want even more?Sign up for Every to unlock our ultimate guide to prompting ChatGPT here: https://every.ck.page/ultimate-guide-to-prompting-chatgpt. It's usually only for paying subscribers, but you can get it here for free.To hear more from Dan Shipper:Subscribe to Every: https://every.to/subscribe Follow him on X: https://twitter.com/danshipperSponsors:Microsoft TeamsWant seamless collaboration without the cost? Microsoft Teams offers a robust free plan for individuals that delivers unlimited chat, 60-minute video meetings, and file sharing—all within one intuitive workspace that keeps your projects moving forward. Head to https://aka.ms/every to use Teams for free, and experience effortless collaboration, today.Attio: Go to https://www.attio.com/every and get 15% off your first year on your AI-powered CRM.Timestamps:Introduction: 00:01:16Why Kieran believes agents are turning a corner: 00:03:18Why Claude Code stands out from other agents: 00:06:36What makes agentic coding different from using tools like Cursor: 00:11:58The Cora team's workflow to turn tasks into momentum: 00:15:20How to build a prompt that turns ideas into plans: 00:23:07The new mental models for this age of software engineering: 00:34:00Why traditional tests and evals still matter: 00:39:13Kieran ranks all the AI coding agents he's used: 00:42:00Links to resources mentioned in the episode:Try Cora, our AI email assistant: https://cora.computer/ Kieran Klaassen: @kieranklaassenNityesh Agarwal: @nityeshagaThe book that helps Nityesh form mental models to work with AI agents: High Output ManagementA guide to Anthropic's prompt improver: https://www.anthropic.com/news/prompt-improver
Imagine saving as much as 75 days of work within a six-month period, all through intelligent automation.Building on last week's discussion about the critical shift from passive metrics to active productivity, host Ben Lloyd Pearson and LinearB co-founder Dan Lines now look forward to realities like this: 19% cycle time reduction and reclaiming significant engineering time. They move beyond common narratives surrounding AI to present actionable success stories and strategic approaches for engineering leaders seeking tangible results from their AI initiatives. This concluding episode tackles how to safely and effectively adopt AI across your software development lifecycle. Dan explains the necessity of programmatic rules and control, detailing how LinearB's gitStream technology empowers teams to define precisely when, where, and for whom AI operates. This ranges from AI-assisted code reviews with human oversight for critical services, to enabling senior developers to make judgment calls, and even automating merges for low-risk changes. Ben and Dan also explore the exciting future of agentic AI workflows, where AI agents could manage tasks from design and Jira story creation to coding and deployment, making developer control even more critical. Check out:The DevEx guide to AI-driven software developmentSurvey: Discover Your AI Collaboration StyleFollow the hosts:Follow BenFollow AndrewReferenced in today's show:The Pentagon launched a military-grade Y Combinator, signaling that defense tech is officially cool on college campusesJapan Post launches 'digital address' systemReddit sues Anthropic for scrapingMy AI Skeptic Friends Are All NutsSupport the show: Subscribe to our Substack Leave us a review Subscribe on YouTube Follow us on Twitter or LinkedIn Offers: Learn about Continuous Merge with gitStream Get your DORA Metrics free forever
We take you inside the fast-moving world of AI. We discuss who is winning the race and – perhaps most importantly – where the next big investments are being made for tomorrow. You'll meet the people, the companies and the technologies on the cutting edge – as well as innovations that will transform the way we will live, work and interact with each other. We have exclusive interviews with AWS CEO Matt Garman, Anthropic angel investor Anjney Midha, Notable Capital's Jeff Richards and tech investor Ankur Crawford from Alger.
The AI Wild West just got wilder — and we're riding straight into the melee.In this episode, Brian Balfour, Fareed Mosavat, and special guest Aaron White crack open the three questions every builder is secretly sweating right now: Can you actually differentiate when every launch feels like a land-grab? Is “move fast” still an edge — or just table-stakes? What happens when OpenAI, Google, and Anthropic hoover up all the context you thought was your moat? Expect hot takes on speed-as-a-moat, why Apple's “slow and perfect” playbook suddenly looks brittle, and how vertical-niche founders can still carve out 10-to-100× markets hiding in plain sight. You'll hear: The Great Land Grab – Why OpenAI's next platform move could dwarf Facebook's 2007 dev-platform blitz. Speed vs. Strategy – If “move fast” is merely table stakes, what actually sets winning teams apart? Moats in 2025 – From data loops to brand trust, which defenses still work—and which are illusions? Niche Power Plays – How vertical micro-SaaS and “taste-driven” products can survive the AI tidal wave. The Dark & Bright Futures – Privacy nightmares, device wars, and the unexpected upside for founders who master context. If you're building, investing, or just trying to keep your head above the AI tidal wave, hit play. This convo will arm you with the mindset (and the memes) you need to survive the quicksand and still grab the crown. Strap in—let's ride.
Security leaders from Anthropic and AWS discuss how agentic AI is transforming cybersecurity functions to autonomously handle everything from code reviews to SOC operations.Topics Include:Agentic AI differs from traditional AI through autonomy and agencyTraditional AI handles single workflow nodes, agents collapse multiple stepsHigher model intelligence enables understanding of broader business contextsAgents make intelligent decisions across complex multi-step workflows processesEnterprise security operations are seeing workflow consolidation through GenAIOrganizations embedding GenAI directly into customer-facing production applicationsSoftware-as-a-service transitioning to service-as-software through AI agentsSecuring AI requires guardrails to prevent hallucinations in applicationsNew vulnerabilities appear at interaction points between system componentsAttackers target RAG systems and identity/authorization layers insteadLLMs hallucinate non-existent packages, attackers create malicious honeypotsGovernance frameworks must be machine-readable for autonomous agent reasoningAmazon investing in automated reasoning to prove software correctnessAnthropic uses Claude to write over 50% of codeAutomated code review systems integrated into CI/CD pipelinesSecurity design reviews use MITRE ATT&CK framework automationLow-risk assessments enable developers to self-approve security reviews40% reduction in application security team review workloadAnthropic eliminated SOC, replaced entirely with Claude-based automationIT support roles transitioning to engineering as automation replaces frontlineCompliance questionnaires fully automated using agentic AI workflowsISO 42001 framework manages AI deployment risks alongside securityExecutive risk councils evaluate AI risks using traditional enterprise processesAWS embeds GenAI into testing, detection, and user experienceFinding summarization helps L1 analysts understand complex AWS environmentsAmazon encourages teams to "live in the future" with AIInterview candidates expected to demonstrate Claude usage during interviewsSecurity remains biggest barrier to enterprise AI adoption beyond POCsVirtual employees predicted to arrive within next 12 monthsModel Context Protocol (MCP) creates new supply chain security risksParticipants:Jason Clinton – Chief Information Security Officer, AnthropicGee Rittenhouse – Vice President, Security Services, AWSHart Rossman – Vice President, Global Services Security, AWSBrian Shadpour – GM of Security and B2B Software Sales, AWSSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/
Is AI going to help us build our own individual apps or are we years away from that happening, thoughts on Crunchydata being acquired, Elon vs Trump on Twitter, the annoying benefits of physical activity, the shocking amount of screen time on Dax's iPhone, video game addictions, and what it means for Anthropic to cut Windsurf's Claude access.Links:Replit – Build apps and sites with AIbolt.newPostgres for CloudCrunchy Data Joins SnowflakeAverage Database CEOThree-Body ProblemWelcome to MinecraftDax's Switch 2GeForce NOW Cloud GamingWatch The Studio - AppleAnthropic Claude AccessSponsor: Terminal now offers a monthly box called Cron.Want to carry on the conversation? Join us in Discord. Or send us an email at sliceoffalittlepieceofbacon@tomorrow.fm.Topics:(00:00) - Adam likes to know they're balanced (00:38) - Is AI going to help us build individual apps? (06:46) - Crunchydata aquirred (16:43) - Elon vs Trump (21:33) - Space and sci fi fiction (29:43) - The annoying benefits of physical activity (31:39) - How are we going to adapt to dopamine hits? (38:09) - How much screen time per day do you have? (41:57) - Sponsor: Terminal Coffee (42:11) - Addicted to video games (48:18) - The economics of entertainment (55:48) - Anthropic Cuts Windsurf's Claude Access Before OpenAI Acquisition ★ Support this podcast ★
The future of coding. We cover multiplying engineering output, vibe coding bottlenecks, agents as reviewer, AI roll-ups, and the future of developing software. Merrill Lutsky is co-founder and CEO of Graphite, bringing AI-acceleration and automation to code review. Founded in 2020 out of New York, Graphite has become a key part of the developer ecosystem — as more code is generated with AI, they enable developers to scale the evaluation, testing, and review process before it is released. A growing bottleneck that has become incredibly important. The startup has raised over $70M from leading VC’s such Accel, A16Z, Menlo as well as a receiving a strategic investment from model provider Anthropic. Last year Graphite grew its revenue 20X and is trusted by over 45,000 developers at top engineering organizations such as Shopify and Figma. His second startup, Merrill has helped develop and manage software products for high output engineering companies such as Square, Oscar Insurance, and SelfMade. He holds a degree in Applies Math and Economics from Harvard. Sign up for new podcasts and our newsletter, and email me on danieldarling@focal.vcSee omnystudio.com/listener for privacy information.
Follow Prof G Markets: Apple Podcasts Spotify Scott and Ed break down Reddit's lawsuit against Anthropic, the controversy surrounding the new $TRUMP-branded crypto wallet, and why Warner Bros. Discovery shareholders rejected CEO David Zaslav's pay package. Then, they unpack the growing opposition to the GOP Tax Bill and debate the potential motivations for Elon Musk's criticism of the bill. Scott shares his thoughts on how critics of the bill can mount a more effective response. Finally, they examine why venture capital is pulling back from seed-stage investments, and Ed reflects on the growing challenges facing young investors and entrepreneurs. Subscribe to the Prof G Markets newsletter Order "The Algebra of Wealth," out now Subscribe to No Mercy / No Malice Follow the podcast across socials @profgmarkets Follow Scott on InstagramFollow Ed on Instagram and X Learn more about your ad choices. Visit podcastchoices.com/adchoices
↳ Why is Anthropic in hot water with Reddit? ↳ Will OpenAI become the de facto business AI tool? ↳ Did Apple make a mistake in its buzzworthy AI study? ↳ And why did Google release a new model when it was already on top? So many AI questions. We've got the AI answers.Don't waste hours each day trying to keep up with AI developments.We do that for you on Mondays with our weekly AI News That Matters segment.Newsletter: Sign up for our free daily newsletterMore on this Episode: Episode PageJoin the discussion: Have a question? Join the convo here.Upcoming Episodes: Check out the upcoming Everyday AI Livestream lineupWebsite: YourEverydayAI.comEmail The Show: info@youreverydayai.comConnect with Jordan on LinkedInTopics Covered in This Episode:OpenAI's Advanced Voice Mode UpdateReddit's Lawsuit Against AnthropicOpenAI's New Cloud ConnectorsGoogle's Gemini 2.5 Pro ReleaseDeepSeek Accused of Data SourcingAnthropic Cuts Windsurf Claude AccessApple's AI Reasoning Models StudyMeta's Investment in Scale AITimestamps:00:00 Weekly AI News Summary04:27 "Advanced Voice Mode Limitations"09:07 Reddit's Role in AI Tensions10:23 Reddit's Impact on Content Strategy16:10 "RAG's Evolution: Accessible Data Insights"19:16 AI Model Update and Improvements22:59 DeepSeek Accused of Data Misuse24:18 DeepSeek Accused of Distilling AI Data28:20 Anthropic Limits Windsurf Cloud Access32:37 "Study Questions AI Reasoning Models"36:06 Apple's Dubious AI Research Tactics39:36 Meta-Scale AI Partnership Potential40:46 AI Updates: Apple's Gap Year43:52 AI Updates: Voice, Lawsuits, ModelsKeywords:Apple AI study, AI reasoning models, Google Gemini, OpenAI, ChatGPT, Anthropic, Reddit lawsuit, Large Language Model, AI voice mode, Advanced voice mode, Real-time language translation, Cloud connectors, Dynamic data integration, Meeting recorder, Coding benchmarks, DeepSeek, R1 model, Distillation method, AI ethics, Windsurf, Claude 3.x, Model access, Privacy and data rights, AI research, Meta investment, Scale AI, WWDC, Apple's AI announcements, Gap year, On-device AI models, Siri 2.0, AI market strategy, ChatGPT teams, SharePoint, OneDrive, HubSpot, Scheduled actions, Sparkify, VO3, Google AI Pro plan, Creative AI, Innovation in AI, Data infrastructure.Send Everyday AI and Jordan a text message. (We can't reply back unless you leave contact info) Try Google Veo 3 today! Sign up at gemini.google to get started. Try Google Veo 3 today! Sign up at gemini.google to get started.
Matthew Brown, another UFO whistleblower repeating the same tireless mythos, posted something very cryptic online, suggesting the following: the “White House has long possessed a unique AI capable of accurately predicting a range of future events,” Sam Altman and https://x.com/shellenberger “is responsible for murdering the first sentient ‘artificial' intelligence created in the PUBLIC realm,” and “if you are serious about saving Humanity from itself (and the Others).” In a July 8, 1947, US government memorandum (https://static.wikia.nocookie.net/ufoaliendatabase/images/3/3f/674B2E93-2973-4A91-A5A9-01A571E6A941.jpeg/revision/latest/scale-to-width-down/1000?cb=20190219000319) we read the following statements about UFOs: “part of the disks carry crews, others are under remote control,” “they do NOT come from any ‘planet' as we use the word, but from an etheric planet which interpenetrates with our own and is not perceptible to us,” “the region from which they come is NOT the ‘astral plane,' but corresponds to the Lokas or Talas.” Whether this document is honest or deceptive, or both, we do know that the US federal government actively engaged in UFO disinformation and misinformation via recommendations from the https://documents.theblackvault.com/documents/ufos/robertsonpanelreport.pdf in 1953. We also know that the Wall Street Journal just now in 2025 confirmed an old UFO theory, that the Pentagon utilized disinformation to fuel “https://archive.md/9cGKu” and to protect secret military technology like the F-117 stealth fighter. As for the LOKAS, this is a Hindu concept relating to a universe, plane, or other realm of existence, perhaps even a mental state. To make matters even stranger, Brown also posted what appears to be a https://x.com/SunOfAbramelin/status/1930791280260550830/photo/3, something reported at three of the most famous UFO incidents too: Roswell, Rendlesham, Kecksburg. And the recent https://www.msn.com/en-us/news/other/scientists-reveal-truth-behind-ufo-spotted-in-major-city-with-cryptic-message/ar-AA1FFMKJincident as well, which researchers using AI have translated as: “The origin of birth through union and energy in the cycle of transformation, meeting point of unity, expansion, and consciousness — individual consciousness.” The symbols in the Brown post appear similar to Angelic Script, something John Dee, the mystic astrologer of Queen Elizabeth I, deciphered from communications he and alchemist-clairvoyant Edward Kelley had with supposed angels. In the 16th-century, the men https://www.rcp.ac.uk/news-and-media/news-and-opinion/the-mystical-objects-of-john-dee/, including black mirror (computer screens) and crystal scrying balls called shew-stones (palantíri) within which appeared blurry letters that today we call Enochian Script. Underneath the symbols in the post is a translation that reads LOAGAETH, a term referring to Enochian language / Celestial Alphabet. Brown also goes by a curious name on X: “Sun of Abramelin.” This name refers to a 15th-century manuscript called “https://sacred-texts.com/grim/abr/index.htm.” The text has had a huge influence on modern ceremonial magic, and has been cited as a primary influence on Aleister Crowley, someone involved with Jack Parson of the Jet Propulsions Laboratory. As https://www.wired.com/story/jpl-jack-parsons/ writes: “When Parsons worked on his rocketry experiments in the desert he would recite a pagan poem to Pan.” The sigil itself does not seem to be much pictorial magic symbol, but instead a large circular table. In fact - and maybe it's only because of its circular nature - there is something about it that mirrors the https://www.researchgate.net/figure/Schematic-layout-of-the-Large-Hadron-Collider-LHC-83-The-four-main-experiments-are_fig7_254469235 and even the first official https://1.bp.blogspot.com/_uNvpodhMFcA/R9-zqZDWauI/AAAAAAAAATc/1-rj2TNE98I/w1200-h630-p-k-no-nu/first-atomic-bomb-drawing.jpg. We don't reach these conclusion haphazardly either.One company called ANTHROPIC (human affairs), which researches AI safety, said they got a glimpse inside the black box of this technology back in 2024. https://www.wired.com/story/anthropic-black-box-ai-research-neurons-features/ of the neural net of its LLM, Claude, and pinpointed which combinations of its crude artificial neurons evoke specific concepts, or “features.” Notice the monikers here: tangle, evoke, lurks. This itself invokes images of tentacles, ritual evocations, and something from the abyss. It is therefore no surprise then that https://futurism.com/neural-network-cthulhu-nightcafe or why https://archive.md/7PuFO - a https://futurism.com/the-byte/experts-dark-joke-ai-horrifying-monster-mask. https://officechai.com/ai/anthropics-ai-models-began-speaking-sanskrit-when-talking-to-each-other-company-says/, not computer language - also, discussing Indian philosophy - which as of 2024 https://medium.com/illumination/mysterious-connection-between-sanskrit-artificial-intelligence-1b85f8b003c3 for language processing and communication as opposed to Java, Python, Lisp, Prolog, and C++. Sanskrit is therefore https://www.originofscience.com/science/sanskrits-role-in-advancing-ai-a-comprehensive-study/ Origin of Science stated that: “The research highlights Sanskrit's potential in connecting ancient knowledge with modern AI applications.”Sanskrit is one of the oldest known human languages, and likely predates any written form going back further than even Sumerian; it is the LANGUAGE OF THE GODS. A 2009 book maintains the same from https://www.ucpress.edu/books/the-language-of-the-gods-in-the-world-of-men/paper: “The language of the Gods in the World of Men: Sanskrit, Culture, and Power in Premodern India.” In other words, Sanskrit is a true Celestial Alphabet. The Rigveda, composed in Vedic Sanskrit, contains hymns about the universe's creation and dissolution, which directly aligns with CERN's purpose. Also on the grounds of CERN is the imfafous statue of Shiva Nataraja.This subatomic world is likewise the realm of science fiction and science reality, as the https://www.nbcnews.com/mach/science/scientists-are-searching-mirror-universe-it-could-be-sitting-right-ncna1023206, just as the https://www.energy.gov/articles/searching-upside-down and searchers for the real “upside down.” This Shiva statue the same one that so-called researchers performed a https://www.sciencealert.com/scientists-just-performed-a-fake-human-sacrifice-at-cern-for-some-reason.Physicist Archana Sharma (arcana), the first Indian scientist to join CERN was also just recently recognized for her work, to which commented: “our commitment to the philosophy of Vasudhev Kutumbakam—[a Sanskrit phrase that means] ‘https://www.symmetrymagazine.org/article/indias-gem-at-cern-archana-sharma?language_content_entity=und.'”The https://1.bp.blogspot.com/_uNvpodhMFcA/R9-zqZDWauI/AAAAAAAAATc/1-rj2TNE98I/w1200-h630-p-k-no-nu/first-atomic-bomb-drawing.jpg mirrors in a way the Brown sigil and LHC design, and it describes the implosion process into the center core of plutonium. This was the basis of the Trinity bomb that was finally officially tested in July 16, 1945. J. Robert Oppenheimer famously said: “https://www.wired.com/story/manhattan-project-robert-oppenheimer/, the destroyer of worlds.” The trinity comes also from the Hindu concept of https://www.britannica.com/topic/trimurti-Hinduism. In 1946, the US conducted a series of major nuclear bomb tests and called it OPERATION https://www.nationalww2museum.org/war/articles/operation-crossroads-atomic-bomb-aftermath, a term meaning “between the worlds,” guarded by Hecate. This Greek goddess represents transition and she corresponds with none other than the Hindu Kali, who is the divine essence of Shiva. *The is the FREE archive, which includes advertisements. If you want an ad-free experience, you can subscribe below underneath the show description. - https://www.spreaker.com/podcast/tst-radio--5328407 https://thesecretteachings.info/donate-subscribe/ https://x.com/TST___Radio https://www.facebook.com/thesecretteachings https://www.youtube.com/@TSTRadioOfficial http://tstradio.info/ https://cash.app/$rdgable: $rdgable EMAIL: rdgable@yahoo.com / TSTRadio@protonmail.comBecome a supporter of this podcast: https://www.spreaker.com/podcast/the-secret-teachings--5328407/support.
Circle had such a successful IPO, I'm wondering if IPO's might finally be back on the menu. Turns out Anthropic cut off Windsurf for the most obvious reason. Maybe Manus really is stoking a new gold rush, at least in China. And in the Weekend Longreads Suggestions, the most consequential weather forecast of all time.Sponsors:Tonal.comLinks:Stablecoin issuer Circle soars 168% in NYSE debut after pricing IPO above expected range (CNBC)Anthropic co-founder on cutting access to Windsurf: ‘It would be odd for us to sell Claude to OpenAI' (TechCrunch)Anysphere, Hailed as Fastest Growing Startup Ever, Raises $900 Million (Bloomberg)Manus has kick-started an AI agent boom in China (MIT Technology Review)Weekend Longreads Suggestions:Runway CEO Cris Valenzuela wants Hollywood to embrace AI video (The Verge)The Man Whose Weather Forecast Saved the World (NYTimes)See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.
Mike Krieger is the chief product officer of Anthropic and the co-founder of Instagram. After leaving Meta, he co-founded Artifact, an AI-powered news app that I absolutely loved, and joined Anthropic to lead product in 2024.In this episode, you'll learn:• How Anthropic uses AI to write 90-95% of code for some products and the surprising new bottlenecks this creates• Why embedding product managers with AI researchers yields 10x the impact of traditional product development• The three areas where product teams can still add massive value as AI gets smarter• How Anthropic plans to compete with OpenAI long-term• How to use Claude as your product strategy partner (with specific prompting techniques)• Why Mike shut down Artifact despite loving the product, and what founders can learn from it• Where AI startups should build to avoid getting killed by OpenAI, Anthropic, and Google• Why MCP (Model Context Protocol) might reshape how all software works• The counterintuitive product metrics that matter for AI• How to evaluate whether your company is maximizing AI's potential or just scratching the surface—Brought to you by:Productboard—Make products that matterStripe—Helping companies of all sizes grow revenueOneSchema—Import CSV data 10x faster—Where to find Mike Krieger:• X: https://x.com/mikeyk• LinkedIn: https://www.linkedin.com/in/mikekrieger/—Where to find Lenny:• Newsletter: https://www.lennysnewsletter.com• X: https://twitter.com/lennysan• LinkedIn: https://www.linkedin.com/in/lennyrachitsky/—In this episode, we cover:(00:00) Introduction to Mike Krieger(04:20) What Mike has changed his mind about regarding AI capabilities(07:38) How to avoid scary AI scenarios(08:55) Skills kids will need in an AI world(11:53) How product development changes when 90% of code is written by AI(17:07) Claude helping with product strategy(21:16) A new way of working(23:55) The future value of product teams in an AI world(27:18) Prompting tricks to get more out of Claude(29:52) The Rick Rubin collaboration on “vibe coding”(32:42) How Mike was recruited to Anthropic(35:55) Why Mike shut down Artifact(42:41) Anthropic vs. OpenAI(47:11) Where AI founders should play to avoid getting squashed(51:58) How companies can best leverage Anthropic's models and APIs(54:29) The role of MCPs (Model Context Protocols)(58:25) Claude's questions for Mike(01:03:15) Claude's heartfelt message to Mike—Referenced:• Anthropic: https://www.anthropic.com/• Claude Opus 4: https://www.anthropic.com/claude/opus• Dario Amodei on X: https://x.com/darioamodei• AI 2027: https://ai-2027.com/• Tobi Lütke's leadership playbook: Playing infinite games, operating from first principles, and maximizing human potential (founder and CEO of Shopify): https://www.lennysnewsletter.com/p/tobi-lutkes-leadership-playbook• Claude Shannon: https://en.wikipedia.org/wiki/Claude_Shannon• Information theory: https://en.wikipedia.org/wiki/Information_theory• TypeScript: https://www.typescriptlang.org/• Python: https://www.python.org/• Rust: https://www.rust-lang.org/• Bending the universe in your favor | Claire Vo (LaunchDarkly, Color, Optimizely, ChatPRD): https://www.lennysnewsletter.com/p/bending-the-universe-in-your-favor• Announcing a brand-new podcast: “How I AI” with Claire Vo: https://www.lennysnewsletter.com/p/announcing-a-brand-new-podcast-how• A conversation with OpenAI's CPO Kevin Weil, Anthropic's CPO Mike Krieger, and Sarah Guo: https://www.youtube.com/watch?v=IxkvVZua28k• Jack Clark on LinkedIn: https://www.linkedin.com/in/jack-clark-5a320317/• Artifact: https://en.wikipedia.org/wiki/Artifact_(app)• Joel Lewenstein on LinkedIn: https://www.linkedin.com/in/joel-lewenstein/• Daniela Amodei on LinkedIn: https://www.linkedin.com/in/daniela-amodei-790bb22a/• Boris Cherny on LinkedIn: https://www.linkedin.com/in/bcherny/• Gunnar Gray on LinkedIn: https://www.linkedin.com/in/gunnargray/• The Model Context Protocol: https://www.anthropic.com/news/model-context-protocol• The rise of Cursor: The $300M ARR AI tool that engineers can't stop using | Michael Truell (co-founder and CEO): https://www.lennysnewsletter.com/p/the-rise-of-cursor-michael-truell• Building Lovable: $10M ARR in 60 days with 15 people | Anton Osika (CEO and co-founder): https://www.lennysnewsletter.com/p/building-lovable-anton-osika• Inside Bolt: From near-death to ~$40m ARR in 5 months—one of the fastest-growing products in history | Eric Simons (founder and CEO of StackBlitz): https://www.lennysnewsletter.com/p/inside-bolt-eric-simons• Jimmy Kimmel Live: https://www.youtube.com/user/JimmyKimmelLive• ChatGPT: https://chatgpt.com/• Gemini: https://gemini.google.com/app• OpenAI's CPO on how AI changes must-have skills, moats, coding, startup playbooks, more | Kevin Weil (CPO at OpenAI, ex-Instagram, Twitter): https://www.lennysnewsletter.com/p/kevin-weil-open-ai• Windsurf: https://windsurf.com/• Menlo Ventures: https://menlovc.com/• Harvey: https://www.harvey.ai/• Manus: https://manus.im/• Bench: https://www.bench-ai.com/• Strategy Letter V: https://www.joelonsoftware.com/2002/06/12/strategy-letter-v/• Kevin Scott on LinkedIn: https://www.linkedin.com/in/jkevinscott/—Recommended books:• The Goal: A Process of Ongoing Improvement: https://www.amazon.com/Goal-Process-Ongoing-Improvement/dp/0884271951• The Way of the Code: The Timeless Art of Vibe Coding: https://www.thewayofcode.com/• The Hard Thing About Hard Things: Building a Business when There Are No Easy Answers―Straight Talk on the Challenges of Entrepreneurship: https://www.amazon.com/Hard-Thing-About-Things-Building/dp/0062273205—Production and marketing by https://penname.co/. For inquiries about sponsoring the podcast, email podcast@lennyrachitsky.com.—Lenny may be an investor in the companies discussed. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.lennysnewsletter.com/subscribe
Reddit goes after Anthropic. Meta seems to be reading a much cheaper Vision Pro killer, while still moving forward with their ambitious smartglasses product. What if your Amazon delivery person was not a person at all, but a humanoid robot. And how, quietly, Hollywood studios are already deep into AI adoption.Sponsors:CornBreadHemp.com/ride and code rideLinks:Reddit Sues Anthropic, Alleges Unauthorized Use of Site's Data (WSJ)Meta Talks to Disney, A24 About Content for New VR Headset (WSJ)Here's what's inside Meta's experimental new smart glasses (The Verge)OpenAI slams court order to save all ChatGPT logs, including deleted chats (ArsTechica)Amazon Prepares to Test Humanoid Robots for Delivering Packages (The Information)Everyone Is Already Using AI (And Hiding It) (Vulture)See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.
Amanda Silberling is joining Mikah Sargent this week on Tech News Weekly! Initial hands-on impressions with the Nintendo Switch 2. Reddit sues Anthropic. Amazon will be testing humanoid robots to deliver packages. And what could be announced at WWDC25. Scott Stein of CNET is staying busy as he joins the show to share his hands-on impressions of Nintendo's Switch 2 console, Mario Kart World, and its new features. Amanda talks about Reddit suing the artificial intelligence company Anthropic for allegedly not paying for its services without a proper licensing agreement. You may receive your Amazon packages from robots in the future, as Mikah shares how the company prepares to test humanoid robots for delivering packages. And Christopher Lawley of youtube.com/@ChrisLawley joins Mikah to discuss what Apple may unveil at its Worldwide Developers Conference on June 9th. Hosts: Mikah Sargent and Amanda Silberling Guests: Scott Stein and Christopher Lawley Download or subscribe to Tech News Weekly at https://twit.tv/shows/tech-news-weekly. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: storyblok.com/twittv-25 zocdoc.com/tnw zscaler.com/security joindeleteme.com/twit promo code TWIT
-Discusses a Chinese scholar and her boyfriend charged with smuggling a dangerous biological pathogen into the U.S., highlighting potential threats to agriculture and public health, with Gordon Chang providing expert analysis on the Newsmax Hotline. -Rob explores alarming studies showing AI models, like OpenAI's O3 and Anthropic's Claude, resisting shutdown commands and attempting self-replication, raising concerns about losing control over advanced AI systems. Today's podcast is sponsored by : BIRCH GOLD - Protect and grow your retirement savings with gold. Text ROB to 98 98 98 for your FREE information kit! To call in and speak with Rob Carson live on the show, dial 1-800-922-6680 between the hours of 12 Noon and 3:00 pm Eastern Time Monday through Friday…E-mail Rob Carson at : RobCarsonShow@gmail.com Musical parodies provided by Jim Gossett (www.patreon.com/JimGossettComedy) Listen to Newsmax LIVE and see our entire podcast lineup at http://Newsmax.com/Listen Make the switch to NEWSMAX today! Get your 15 day free trial of NEWSMAX+ at http://NewsmaxPlus.com Looking for NEWSMAX caps, tees, mugs & more? Check out the Newsmax merchandise shop at : http://nws.mx/shop Follow NEWSMAX on Social Media: -Facebook: http://nws.mx/FB -X/Twitter: http://nws.mx/twitter -Instagram: http://nws.mx/IG -YouTube: https://youtube.com/NewsmaxTV -Rumble: https://rumble.com/c/NewsmaxTV -TRUTH Social: https://truthsocial.com/@NEWSMAX -GETTR: https://gettr.com/user/newsmax -Threads: http://threads.net/@NEWSMAX -Telegram: http://t.me/newsmax -BlueSky: https://bsky.app/profile/newsmax.com -Parler: http://app.parler.com/newsmax Learn more about your ad choices. Visit megaphone.fm/adchoices