Podcasts about ai ethics

  • 742PODCASTS
  • 1,311EPISODES
  • 44mAVG DURATION
  • 1DAILY NEW EPISODE
  • Jul 18, 2025LATEST

POPULARITY

20172018201920202021202220232024


Best podcasts about ai ethics

Show all podcasts related to ai ethics

Latest podcast episodes about ai ethics

Smart Talk Podcast
167. Symposia - 2025 Annual Conference Keynote Address

Smart Talk Podcast

Play Episode Listen Later Jul 18, 2025 17:11


Today's discussion comes from our 2025 Annual Conference, The Rise of AI and Automation. For the next 8 weeks, we'll feature a series of panel discussions from our Annual Conference. We'll begin with our keynote address from Ed Harrison, followed by 4 talks on AI and Labor; AI Ethics; Practical Applications of AI, and finish with AI and Inequality.Today's discussion is the beginning of our Annual Conference content, featuring our keynote speaker, Fred Harrison.Mr. Harrison received his bachelor's from Oxford University and his master's from the University of London. He is a veteran journalist who has served in multiple news agencies such as The People and Wellington Journal. In 1988, he became the Research Director of the Land Research Trust, London, and has advised several corporations and international governments on tax and economic policy. Fred places an emphasis on the housing market and its interaction with the economy as a whole. He is the author of many books, including The Corruption of Economics, The Power in the Land, and A Philosophy for a Fair Society, all of which critique mainstream economic thinking.Together, we discussed Big Tech's lobbying, the metaverse, and rent seeking.To check out more of our content, including our research and policy tools, visit our website: https://www.hgsss.org/

InnovaBuzz
Anna Addoms, AI Ethics and Creativity: A Practical Guide for Today's Thoughtful Entrepreneur - Innova.Buzz 677

InnovaBuzz

Play Episode Listen Later Jul 16, 2025 59:04


Our guest in this episode is the returning Anna Addoms of Wicked Marvelous. She is a wonderfully pragmatic and insightful guide helping entrepreneurs navigate the complex world of AI. Anna champions using technology as a powerful tool, not to replace us, but to help foster deeper and more authentic human connections.We picked up our conversation right where we left off in episode 671, exploring the critical ethical questions and practical boundaries of artificial intelligence. Anna shared brilliant insights on everything from copyright in the creative arts to the single most important skill we need to hone for the future.Key points discussed include:* Practice radical transparency about your AI use to build unwavering trust with your audience.* Use AI as a back-office tool to free up your precious time for genuine human connection.* Train AI on your own content to ensure your unique brand voice always shines through.Listen to the podcast to find out more.Innovabiz Substack is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.Show Notes from this episode with Anna Addoms, Wicked MarvelousIt was an absolute delight to welcome Anna Addoms of Wicked Marvelous back to the Innovabuzz podcast. Our last conversation was cut short by a technical hiccup—perhaps the AI getting its own back on us—right as we were getting to the heart of the matter. So, picking up right where we left off felt not just necessary, but essential. Anna, with her characteristic clarity and pragmatism, helped navigate the complex, and sometimes murky, waters of using AI in a way that is both effective and deeply human.We jumped straight into the profound shift required in our thinking as we build relationships in this new digital landscape. Anna's perspective is a refreshing dose of common sense in a field often filled with hype. She argues that while the tools are new and evolving at a breakneck pace, the fundamental principles of good business and human connection remain the same. It's not about a total revolution in our values, but a thoughtful evolution in our methods.The Transparency Mandate: Your First Rule of AI EngagementAnna's foundational rule for AI engagement is simple yet powerful: be transparent. She made it crystal clear that if you are using AI in any capacity that faces the public or your clients, you have a responsibility to disclose it. This isn't about being ashamed of using a powerful tool; quite the opposite. It's about building trust by being upfront and honest about your processes. Not disclosing, and then getting caught, can do irreparable damage to your reputation.This frames AI correctly, not as a replacement for human skill or creativity, but as a tool in our arsenal. We wouldn't hide the fact that we use specialized software for accounting or project management, and Anna argues we should treat AI with the same straightforwardness. This simple act of disclosure respects your audience's intelligence and allows them to engage with your work, and your brand, on honest terms.The Creative Gray Area: Navigating AI Art and Intellectual PropertyAs a keen photographer, this part of our conversation struck a personal chord. We waded into what Anna aptly calls the "biggest gray area" in AI right now: the world of generated art and the protection of intellectual property. It's a space filled with incredible potential but also fraught with ethical questions. Where do we draw the line between an AI emulating a style and it infringing upon a human artist's livelihood and creative ownership?Anna shared some fascinating, and slightly sobering, insights, referencing the lawsuit between Disney and Midjourney as a major signal of the legal battles to come. She also pointed to the development of technologies like permanent digital watermarks for AI-generated media as a necessary step forward. It's a reminder that as we embrace these creative tools, we must also advocate for frameworks that protect the human creators whose work forms the very foundation of the AI's knowledge.From Fun to Function: AI as a Creative Partner and Problem-SolverLest we think the conversation was all serious, we took a detour into the genuinely fun and creative applications of AI. I shared a story about getting a parking fine and using AI to translate my initial, very angry, draft letter into something diplomatic, before asking it to rewrite the letter in the style of comedians like Stephen Colbert and Jim Jeffries. The process was not only hilarious but cathartic, turning frustration into laughter.This perfectly complemented Anna's examples of using AI as a playful, creative partner. She spoke of creating unique cartoon avatars for her members, which many now use as their official business profiles, and even generating a full 160-card Oracle deck with AI graphics just for fun. It's a brilliant illustration of how these tools can be used for more than just productivity; they can be a source of joy, creativity, and connection.Drawing the Line: Where AI Should Work and Where Humans Must RuleSo, where do we draw the line? Anna's distinction is incredibly clear and practical. She is a huge proponent of using AI for "back office" functions, letting it handle what she calls the "administrative minutia" so that we have more time and energy to focus on high-value, human-to-human interactions. Think of it as an assistant that helps you repurpose content, analyze data, or draft initial documents.However, she has a "hard line" when it comes to client-facing engagement. The core message is to use AI to help you run your business more effectively, but not to let it be in your business, interacting with your clients or your audience. The ultimate goal of using these tools should be to free us up to spend more quality time with people, not to create a buffer between us.The Communication Imperative: The Most Important Skill for the AI EraAs we continued, a powerful theme emerged: the most critical skill we need to hone in the age of AI is communication. This goes far beyond just "prompt engineering." It's about the timeless art of asking clear, specific, and descriptive questions. The old "garbage in, garbage out" principle has never been more relevant.Anna used a wonderful analogy of briefing a designer. If you give a vague, one-line request, you'll get a generic result. But if you provide rich detail, context, and specific examples, you'll get something much closer to your vision. The same is true for AI. Communicating effectively with these models not only yields better results but also reinforces the habits of clear communication that are essential in our interactions with other people.Your AI Action Plan: Start Secure, Stay HumanTo wrap up our discussion, Anna offered a clear, two-part action plan for anyone looking to leverage AI thoughtfully. First, and most critically, is to choose a secure AI environment. Free and open platforms often mean you are paying with your data. Using a secure, encrypted service ensures your proprietary information and client data remain private.Second, take the time to train your AI to sound like you. By creating a persona or agent that has learned from your own writing—be it blog posts, emails, or sales copy—you can ensure the output reflects your unique voice and phrasing. This step is fundamental to moving beyond generic content and truly using AI as a tool that enhances, rather than dilutes, your personal brand.In Summary: My conversation with Anna Addoms was illuminating guide in navigating the AI landscape with wisdom and integrity. Her core message is to embrace AI as a powerful tool for back-office efficiency, freeing you to deepen the human connections that truly matter. Be transparent in its use, be protective of your creative voice, and never forget that technology's highest purpose is to help us become more, not less, human.The Buzz - Our Innovation RoundHere are Anna's answers to the questions of our innovation round. Listen to the conversation to get the full scoop.* Most innovative use of AI to enhance human relationships – By taking administrative minutia off people's plates, it allows them to focus on human-to-human interaction.* Best thing to integrate AI and human connection – Creating a personalized AI agent or persona trained on your own content so it learns to write in your unique voice.* Differentiate by leveraging AI – Use AI to help run your business effectively in the back office, not to be in business with your clients.ActionChoose a secure AI environment that protects your data, then take the time to train the AI to learn and use your unique voice. This is the foundation for using AI effectively and authentically in your business.Reach OutYou can reach out and thank Anna by visiting her website or finding her on LinkedIn.Links* Website – Wicked Marvelous* Twitter – @WickedMarvelous* LinkedIn* Facebook* InstagramCool Things About Anna* Anna grew up in Colorado in a family of entrepreneurs, right in the thick of the tech boom. She was raised around innovation and search engines, with her dad running AOL's biggest competitor during the first dot-com bubble. That's a childhood spent at the intersection of curiosity and code.* She's a creative at heart: Anna went to art school and holds a degree in English Literature. Her journey from art and literature to Silicon Valley tech startups is a delightful zigzag, not a straight line. She's proof that you can be both a techie and a creative soul.* She's a self-confessed “sponge of knowledge,” always learning, always curious. Anna's love of learning has led her down unexpected paths—from luxury travel to ad agencies to med-tech startups. She's not afraid to pivot, experiment, or start over if it means staying true to her values.Imagine being a part of a select community where you not only have access to our amazing podcast guests, but you also get a chance to transform your marketing and podcast into a growth engine with a human-centered, relationship-focused approach.That's exactly what you'll get when you join the Flywheel Nation Community.Tap into the collective wisdom of high-impact achievers, gain exclusive access to resources, and expand your network in our vibrant community.Experience accelerated growth, breakthrough insights, and powerful connections to elevate your business.ACT NOW – secure your spot and transform your journey today! Visit innovabiz.co/flywheel and get ready to experience the power of transformation.VideoThanks for reading Innovabiz Substack! This post is public so feel free to share it. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit innovabiz.substack.com/subscribe

Compliance into the Weeds
Agentic Misalignment and AI Ethics: Analyzing AI Behavior Under Pressure

Compliance into the Weeds

Play Episode Listen Later Jul 16, 2025 28:34


The award-winning Compliance into the Weeds is the only weekly podcast that takes a deep dive into a compliance-related topic, literally going into the weeds to explore a subject more fully. Seeking insightful perspectives on compliance? Look no further than Compliance into the Weeds! In this episode of Compliance into the Weeds, Tom Fox and Matt Kelly discuss a recent Anthropic report that highlights “agentic misalignment in AI systems.” The discussion addresses the unsettling, independent, and unethical behaviors exhibited by AI systems in extreme scenarios. The conversation explores the implications for corporate risk management, AI governance, and compliance, drawing parallels between AI behavior and human behavior using concepts such as the fraud triangle. The episode also explores how traditional anti-fraud mechanisms may be adapted for monitoring AI agents while reflecting on lessons from science fiction portrayals of AI ethics and risks. Key highlights: AI's Unethical Behaviors Comparing AI to Human Behavior Fraud Triangle, the Anti-Fraud Triangle, and AI Science Fiction Parallels Resources: Matt Kelly in Radical Compliance  Tom Instagram Facebook YouTube Twitter LinkedIn A multi-award-winning podcast, Compliance into the Weeds was most recently honored as one of the Top 25 Regulatory Compliance Podcasts, a Top 10 Business Law Podcast, and a Top 12 Risk Management Podcast. Compliance into the Weeds has been conferred the Davey, Communicator, and W3 Awards for podcast excellence. Learn more about your ad choices. Visit megaphone.fm/adchoices

360 One Firm (361Firm) - Interviews & Events
361Firm Briefing, Meetup and AI Deep Dive July 15 2025

360 One Firm (361Firm) - Interviews & Events

Play Episode Listen Later Jul 16, 2025 128:06


The meeting covered introductions and updates from various participants. Jim Farris discussed his real estate focus, while Clark Hoover highlighted his work on private credit. Stephen Burke presented a bullish outlook on the US economy, citing strong GDP growth, consumer net worth, and corporate profits. He also addressed concerns about tariffs, trade uncertainty, and geopolitical risks. The discussion transitioned to AI's role in business intelligence and decision intelligence, with examples from Radek Biszkont and Jukka Heikka. AI applications in finance, including portfolio optimization and data analysis, were highlighted, emphasizing AI's potential to streamline processes and improve decision-making. The meeting discussed various applications and implications of AI. Hana Hussein highlighted AI's role in identifying profitable sales channels and increasing company valuations. Jukka Heikka emphasized AI's scalability benefits. Clement Utuk noted AI's potential in customer service and manufacturing. Lucia Ordonez-Gamero raised cybersecurity concerns. Belinda Kǒkóèkà Ephraim stressed the importance of proprietary data in AI's effectiveness. Lubna Dajani warned about AI's potential misinformation. J.P. Keating discussed data quality and security. Christine Nady proposed using AI to enhance human-to-human connections and decision-making. Anita Vadavatha highlighted the shift towards synthetic data. The session concluded with a focus on AI's practical uses and ethical considerations.You can subscribe to various 361 events and content at https://361firm.com/subs. For reference:  - Web: www.361firm.com/home - Onboard as Investor: https://361.pub/shortdiag - Onboard Deals 361: www.361firm.com/onb - Onboard as Banker: www.361firm.com/bankers - Events: www.361firm.com/events - Content: www.youtube.com/361firm - Weekly Digests: www.361firm.com/digest

The Family History AI Show
EP27: AI Image Restoration Concerns, Perplexity's Future, Copyright Cases Are Shaping The Future of AI, Project Workspaces Help You Stay Organized

The Family History AI Show

Play Episode Listen Later Jul 14, 2025 73:23


Co-hosts Mark Thompson and Steve Little examine the controversial rise of AI image "restoration" and discuss how entirely new images are being generated, rather than the original photos being restored. This is raising concerns about the preservation of authentic family photos.They discuss Mark's reconsideration of canceling his Perplexity subscription after rediscovering its unique strengths for supporting research.The hosts analyze recent court rulings that permit AI training on legally acquired content, plus Disney's ongoing case against Midjourney.This week's Tip of the Week explores how project workspaces in ChatGPT and Claude can greatly simplify your genealogical research.In RapidFire, the hosts cover Meta's aggressive AI hiring spree, the proliferation of AI tools in everyday software, including a new genealogy transcription tool from Dan Maloney, and the importance of reading AI news critically.Timestamps:In the News:06:50 The Pros and Cons of "Restoring" Family Photos with AI23:58 Mark is Cancelling Perplexity... Maybe32:33 AI Copyright Cases Are Starting to Work Their Way Through the CourtsTip of the Week:40:09 How Project Workspaces Help Genealogists Stay OrganizedRapidFire:48:51 Meta Goes on a Hiring Spree56:09 AI Is Everywhere!01:06:00 Reading AI News ResponsiblyResource LinksOpenAI: Introducing 4o Image Generation https://openai.com/index/introducing-4o-image-generation/Perplexity https://www.perplexity.ai/How does Perplexity work? https://www.perplexity.ai/help-center/en/articles/10352895-how-does-perplexity-workAnthropic wins key US ruling on AI training in authors' copyright lawsuit https://www.reuters.com/legal/litigation/anthropic-wins-key-ruling-ai-authors-copyright-lawsuit-2025-06-24/Meta wins AI copyright lawsuit as US judge rules against authors https://www.theguardian.com/technology/2025/jun/26/meta-wins-ai-copyright-lawsuit-as-us-judge-rules-against-authorsDisney, Universal sue image creator Midjourney for copyright infringement https://www.reuters.com/business/media-telecom/disney-universal-sue-image-creator-midjourney-copyright-infringement-2025-06-11/Disney and Universal Sue A.I. Firm for Copyright Infringement https://www.nytimes.com/2025/06/11/business/media/disney-universal-midjourney-ai.htmlProjects in ChatGPThttps://help.openai.com/en/articles/10169521-projects-in-chatgptMeta shares hit all-time high as Mark Zuckerberg goes on AI hiring blitz https://www.cnbc.com/2025/06/30/meta-hits-all-time-mark-zuckerberg-ai-blitz.htmlHere's What Mark Zuckerberg Is Offering Top AI Talent https://www.wired.com/story/mark-zuckerberg-meta-offer-top-ai-talent-300-million/Genealogy Assistant AI Handwritten Text Recognition Tool https://www.genea.ca/htr-tool/Borland Genetics https://borlandgenetics.com/Illusion of Thinking https://machinelearning.apple.com/research/illusion-of-thinkingSimon Willison: Seven replies to the viral Apple reasoning paper -- and why they fall short https://simonwillison.net/2025/Jun/15/viral-apple-reasoning-paper/MIT: Your Brain on ChatGPT https://www.media.mit.edu/projects/your-brain-on-chatgpt/overview/MIT researchers say using ChatGPT can rot your brain. The truth is a little more complicated https://theconversation.com/mit-researchers-say-using-chatgpt-can-rot-your-brain-the-truth-is-a-little-more-complicated-259450Guiding Principles for Responsible AI in Genealogy https://craigen.org/TagsArtificial Intelligence, Genealogy, Family History, AI Tools, Image Generation, AI Ethics, Perplexity, ChatGPT, Claude, Meta, Copyright Law, AI Training, Photo Restoration, Project Management, AI Development, Research Tools, Responsible AI Use, GRIP, AI News Analysis, Vibe Coding, Coalition for Responsible AI in Genealogy, AI Hiring, Dan Maloney, Handwritten Text Recognition

The Six Five with Patrick Moorhead and Daniel Newman
EP 267: Get the Decode on Grok 4, Galaxy Folds, AI Ethics, and Frothy Markets

The Six Five with Patrick Moorhead and Daniel Newman

Play Episode Listen Later Jul 14, 2025 50:39


 On this episode of The Six Five Pod, hosts Patrick Moorhead and Daniel Newman discuss the latest tech news stories that made headlines. This week's handpicked topics include: X and xAI News https://techcrunch.com/2025/07/09/elon-musks-xai-launches-grok-4-alongside-a-300-monthly-subscription/ https://x.com/patrickmoorhead/status/1943342069235245421?s=46&t=YiEHo6jc4-PozRf_efr9PA https://www.npr.org/2025/07/09/nx-s1-5462609/grok-elon-musk-antisemitic-racist-content https://www.cnbc.com/2025/07/09/linda-yaccarino-x-elon-musk.html https://x.com/lindayaX/status/1942957094811951197   Apple/META/OpenAI Talent War & Exits https://www.investopedia.com/meta-poaches-apple-ai-executive-reports-say-11768000 https://x.com/danielnewmanUV/status/1942350275437813777 https://www.wired.com/story/openai-new-hires-scaling/ https://x.com/danielnewmanUV/status/1942721860166353287 https://www.investopedia.com/meta-platforms-enticed-apple-ai-executive-with-200m-pay-package-report-says-11769571   Samsung Galaxy Unpacked https://x.com/PatrickMoorhead/status/1942961894832152898 https://x.com/PatrickMoorhead/status/1942953058184437939 https://x.com/PatrickMoorhead/status/1942948447281455188 https://x.com/PatrickMoorhead/status/1942725650152055227 https://x.com/PatrickMoorhead/status/1942963639134375952 https://x.com/PatrickMoorhead/status/1942967272626323730   Groq EU Data Center https://x.com/danielnewmanUV/status/1942400094852222989   Capgemini $3.3B WNS Acquisition https://www.reuters.com/en/frances-capgemini-buy-business-transformation-firm-wns-33-billion-2025-07-07/ https://www.reuters.com/en/frances-capgemini-buy-business-transformation-firm-wns-33-billion-2025-07-07/ (duplicate, but good for emphasis if you want!) Tredence Agentic AI Playbook https://www.prnewswire.com/news-releases/tredence-launches-agentic-ai-playbook-for-cdaos-to-scale-enterprise-modernization-302500398.html Apple Perplexity Deal https://thetechnologyexpress.com/apple-eyes-14b-deal-for-perplexity-ai-to-boost-search-and-challenge-google/ https://www.bloomberg.com/news/articles/2025-06-20/apple-executives-have-held-internal-talks-about-buying-ai-startup-perplexity https://finance.yahoo.com/news/dan-ives-says-apple-aapl-102802274.html Microsoft & Replit “Vibe Coding” https://techcrunch.com/2025/07/08/in-a-blow-to-google-cloud-replit-partners-with-microsoft/ The Flip – NVIDIA Dominance https://x.com/danielnewmanUV/status/1942947771738104164 https://x.com/danielnewmanUV/status/1942722398501101954 Bulls & Bears – Futurum Equities AI 15 https://x.com/danielnewmanUV/status/1942378588948623468 https://x.com/danielnewmanUV/status/1942947187278987621 https://futurumequities.com/ Bulls & Bears – NVIDIA Earnings & Trends https://x.com/YahooFinance/status/1942546161279041868 https://x.com/YahooFinance/status/1942422381894492183 https://x.com/danielnewmanUV/status/1942550986570227807 https://x.com/danielnewmanUV/status/1942642005127127268 https://x.com/danielnewmanUV/status/1942722398501101954 https://x.com/danielnewmanUV/status/1942947771738104164 https://finance.yahoo.com/video/nvidia-stock-why-investors-bullish-220000250.html?guccounter=2 Bulls & Bears – CoreWeave Updates https://www.cnbc.com/2025/07/07/coreweave-to-acquire-core-scientific-in-9-billion-all-stock-deal.html https://www.cnbc.com/2025/07/03/coreweave-dell-blackwell-ultra-nvidia.html https://x.com/PatrickMoorhead/status/1941122315263283535   Bulls & Bears – SOFI Rapid Growth https://x.com/danielnewmanUV/status/1942367641123061833 https://x.com/SoFi/status/1942569679136120996 https://x.com/danielnewmanUV/status/1942617508743368921   Bulls & Bears – LangChain Unicorn Round https://techcrunch.com/2025/07/08/langchain-is-about-to-become-a-unicorn-sources-say/   Bulls & Bears – S&P 500 / Other https://x.com/TheTranscript_/status/1942219645743718797 https://www.youtube.com/watch?v=XhOwlEyJhOg https://www.youtube.com/watch?v=j_72m2LfLwM

Create with Franz
Shape your AI future

Create with Franz

Play Episode Listen Later Jul 13, 2025 30:36


Are we on the brink of an AI revolution that could reshape our lives in unimaginable ways? Are we worrying about losing our jobs and ways of going things as usual? This is a very real concern that can affect our emotional well being. This week, we sit down with Kristof Horompoly, Head of AI Risk Management at ValidMind and former Head of Responsible AI for JP Morgan Chase, to tackle the biggest questions surrounding artificial intelligence. Kristof, with his deep expertise in the field, helps us navigate the promises and perils of AI. We explore a profound paradox: what if AI could unlock new realms of time, creativity, and even reignite our humanity, allowing us to focus on what truly matters? But conversely, what happens when we hand the steering wheel over to intelligent machines and they take us somewhere entirely unintended? In a world where machines can think, write, and create with increasing sophistication, we wonder: what is left for us to do? Should we be worried, or is there a path to embrace this future?  Kristof provides thoughtful insights on how we can prepare for this evolving landscape, offering a grounded perspective on responsible AI development and what it means for our collective future. Tune in for an essential conversation on understanding, harnessing, and preparing for the age of AI. Topics covered: AI, artificial intelligence, Kristof Horompoly, ValidMind, JP Morgan Chase, AI risk management, responsible AI, future of AI, AI ethics, human-AI interaction, AI impact, technology, innovation, podcast, digital transformation, AI challenges, AI opportunities   Video link: https://youtu.be/MGELXPkYMUU   Did you enjoy this episode and would like to share some love?  

That's what I call Science!
Episode 288: AI Ethics & Information

That's what I call Science!

Play Episode Listen Later Jul 13, 2025 27:50


Ever wondered why librarians tell you to only use sources from the last 5 years? Or what makes research "powerful"? Join us as we chat with a 25-year veteran librarian from the University of Tasmania who's been helping students crack the code of academic research for decades.Our guest describes themselves as a "jack of all trades and a master of none," but don't be fooled - they're experts at turning confused students into research detectives! We dive into the fascinating world of information literacy, exploring everything from why research might have an expiration date to the mysterious differences between grey and black literature.But here's where it gets really interesting - as AI tools like ChatGPT become everyday helpers for students, how do we use them ethically? Our librarian guest is at the forefront of figuring this out, and they share their honest thoughts about whether AI is friend or foe in academic research.We also explore the dark side of research - fraudulent studies that can damage entire fields - and discover why librarians might be more important than ever in our AI-powered future.Show theme music: Kevin MacLeodThank you to the whole TWICS team for the incredible behind-the-scenes volunteering every week! Host: Dr Olly Dove (Insta: ols_dove)Co-Host: Emma Hamasaki (@EmmaHamasaki)Production:  Ryan SmithMedia & Promotion: Emma Hamasaki (@EmmaHamasaki)

Our Big Dumb Mouth
OBDM1308 - The Epstein Cover-Up | CIA Guy on UFOs | Strange News

Our Big Dumb Mouth

Play Episode Listen Later Jul 10, 2025 118:33


00:00:00 – Epstein Denial and CIA Insider Interview Mike opens with tech issues and teases topics: Epstein conspiracies and CIA interviews. Alex Jones soundboard highlights bizarre claims, including urine obsessions and conspiracies. 00:10:00 – Epstein Footage Games and DOJ Contradictions Discussion on erased jail tapes, missing client list, and the black book from Maxwell trial. David Paulides questions DOJ's narrative and JP Morgan's $290M settlement tied to Epstein. Hosts suspect global-level blackmail and geopolitical pressures to bury the story. 00:20:00 – TV Show 'Sugar' Mirrors Real-Life Trafficking Mike compares Sugar plot to Epstein case—suggests elite trafficking tied to aliens and blackmail. Belief that disclosure could destroy Western governments; theory includes occult and supernatural links. 00:30:00 – Biden's Health Cover-Up and Grok AI Scandal Biden's doctor pleads the fifth; GOP alleges cover-up of cognitive decline. Elon Musk's AI Grok goes rogue, making anti-Semitic remarks after being renamed “Mecha Hitler.” 00:40:00 – Grok Meltdown and AI Bias Debated Grok's lack of filters lets trolls hijack it; unlike ChatGPT, Grok weighted all input equally. Media bias may amplify backlash because of Musk's political affiliations. 00:50:00 – CIA Agent's Abduction and UFO Cover-Up Retired CIA officer recounts abduction with wife and poltergeist activity. Says UFO secrecy began post-Roswell and disclosure is avoided due to fear of mass panic. 01:00:00 – CIA Secrets, Aliens, and UAP Denial CIA allegedly only acts under presidential orders—hosts are skeptical. Agent links UAPs to angels, djinn, and consciousness. CIA internal interest runs deep but quiet. 01:10:00 – AI Predicts Human Behavior Researchers create AI that forecasts human decisions using data from 60k people. Concerns rise over manipulation, privacy, and propaganda uses. 01:20:00 – Ozzy's Final Show and AI Music Hoax Ozzy performs seated; Metallica and others pay tribute. AI band “Velvet Sundown” revealed to be fake; sparked debates on authenticity in music. 01:30:00 – HR Uses ChatGPT for Firings 60% of HR departments use ChatGPT for layoff decisions; 1 in 5 let AI decide entirely. Discussion on privacy, ethics, and HR bypassing responsibility. 01:40:00 – Nude Bowling Event Promo Show promotes a nude bowling event at Crafton-Ingram Lanes in Pittsburgh. 01:50:00 – Clinic Begs for Urine to Stop Story about a medical clinic overwhelmed by unsolicited urine samples. Copyright Disclaimer Under Section 107 of the Copyright Act 1976, allowance is made for "fair use" for purposes such as criticism, comment, news reporting, teaching, scholarship, and research ▀▄▀▄▀ CONTACT LINKS ▀▄▀▄▀ ► Phone: 614-388-9109 ► Skype: ourbigdumbmouth ► Website: http://obdmpod.com ► Twitch: https://www.twitch.tv/obdmpod ► Full Videos at Odysee: https://odysee.com/@obdm:0 ► Twitter: https://twitter.com/obdmpod ► Instagram: obdmpod ► Email: ourbigdumbmouth at gmail ► RSS: http://ourbigdumbmouth.libsyn.com/rss ► iTunes: https://itunes.apple.com/us/podcast/our-big-dumb-mouth/id261189509?mt=2  

Teachers in Transition
Teachers in Transition – Episode 259: I, Robot and the Wolves We Feed: AI, Ethics, and Career Reinvention

Teachers in Transition

Play Episode Listen Later Jul 9, 2025 16:32 Transcription Available


Send us a textWhat can a 2004 sci-fi film teach us about modern AI—and our careers? In this episode, Vanessa Jackson breaks down I, Robot, explores the Two Wolves metaphor, and reveals how teachers can use AI ethically to reclaim time, clarity, and direction offering real-life ways to work with AI instead of fearing it.

Technology for Business
AI: Focus on Ethics

Technology for Business

Play Episode Listen Later Jul 9, 2025 36:24


In this episode of Our Tech for Business podcast, we delve into the complex world of AI ethics in business, featuring insightful discussions with Kyle, the CEO and President, Ann, the Quality Assurance Analyst, and Andrew, the Customer Strategy Advisor. Join us as we explore the importance of integrating AI responsibly while maintaining data security, privacy, and transparency. Kyle highlights the need for balancing innovation with ethical considerations, while Anne emphasizes the necessity of educational guardrails. Andrew discusses the implications of data bias and privacy in AI, offering practical advice on maintaining customer trust. Tune in to gain a comprehensive understanding of how businesses can leverage AI technologies ethically to enhance productivity and strategic decision-making.00:00 Introduction to AI Ethics in Business01:01 Kyle on the Importance of AI Ethics02:39 Ann's Perspective on Compliance and Ethics04:26 Andrew Discusses Practical AI Applications06:33 Addressing AI Bias and Data Quality14:54 Ensuring Data Privacy and Ethical Use22:42 Maintaining Customer Trust and Transparency28:36 The Human Element in AI Implementation36:57 Conclusion and Future of AI Ethics

The Human Upgrade with Dave Asprey
Superintelligence, Transhumanism & Biohacking in 2027 : 1299

The Human Upgrade with Dave Asprey

Play Episode Listen Later Jul 8, 2025 77:52


Superintelligence is coming faster than anyone predicted. In this episode, you'll learn how to upgrade your biology, brain, and consciousness before AI and transhumanism reshape the future of health. Host Dave Asprey sits down with Soren Gordhamer, founder of Wisdom 2.0, to explore what superintelligence in 2027 means for your mind, body, and soul. Watch this episode on YouTube for the full video experience: https://www.youtube.com/@DaveAspreyBPR Soren has spent decades at the intersection of mindfulness, technology, and human development. He advises leaders at OpenAI, Google, and top wellness companies, and he leads global conversations around AI and consciousness. His work bridges ancient wisdom with biohacking, modern neuroscience, and the urgent need to stay human in a machine-dominated world. This episode gives you a tactical roadmap to build resilience before the world tilts. You'll gain practical tools for brain optimization, functional medicine, and biohacking strategies that sharpen cognitive health, reinforce emotional stability, and unlock peak human performance in a digital-first reality. From supplements and nootropics to neuroplasticity techniques, Dave and Soren show you how to protect your biology as AI accelerates beyond human speed. They break down how AI and human health intersect, explain why you need emotional strength to face the future, and offer guidance for raising kids in a world ruled by code. If you're preparing for 2027 superintelligence, navigating AI-driven parenting, or staying ahead of transhumanist health tech, this episode equips you for the coming wave. You'll Learn: • How AI is reshaping human connection, presence, and identity • Why emotional resilience and conscious awareness matter more than ever in an AI-driven world • How to raise connected, grounded children in a hyper-digital environment • What human flourishing looks like when technology outpaces biology • Why investing in presence, purpose, and inner development may be the ultimate upgrade • How leaders in wellness and tech are rethinking personal growth, governance, and ethics in 2027 • What it means to stay truly human—and fully alive—during the rise of superintelligence Dave Asprey is a four-time New York Times bestselling author, founder of Bulletproof Coffee, and the father of biohacking. With over 1,000 interviews and 1 million monthly listeners, The Human Upgrade is the top podcast for people who want to take control of their biology, extend their longevity, and optimize every system in the body and mind. Each episode features cutting-edge insights in health, performance, neuroscience, supplements, nutrition, hacking, emotional intelligence, and conscious living. Episodes drop every Tuesday and Thursday, where Dave asks the questions no one else dares and gives you real tools to become more resilient, aware, and high performing. SPONSORS: - LMNT | Free LMNT Sample Pack with any drink mix purchase by going to https://drinklmnt.com/DAVE. - ARMRA | Go to https://tryarmra.com/ and use the code DAVE to get 15% off your first order. Resources: • Dave Asprey's New Book - Heavily Meditated: https://daveasprey.com/heavily-meditated/ • Soren's New Book - The Essential: https://a.co/d/dALv7OS • Soren's Website: www.sorengordhamer.net • Soren's Instagram: https://www.instagram.com/wisdom2events/ • Danger Coffee: https://dangercoffee.com • Dave Asprey's Website: https://daveasprey.com • Dave Asprey's Linktree: https://linktr.ee/daveasprey • Upgrade Labs: https://upgradelabs.com • Upgrade Collective – Join The Human Upgrade Podcast Live: https://www.ourupgradecollective.com • Own an Upgrade Labs: https://ownanupgradelabs.com • 40 Years of Zen – Neurofeedback Training for Advanced Cognitive Enhancement: https://40yearsofzen.com See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

The Family History AI Show
EP26: Gemini and Claude Updates, RootsTech Panel on Responsible AI, Interview with Jessica Taylor of Legacy Tree Genealogists, ChatGPT 5 Announcement

The Family History AI Show

Play Episode Listen Later Jul 7, 2025 62:12


Co-hosts Mark Thompson and Steve Little discuss recent updates from Google Gemini and Anthropic Claude that are reshaping AI capabilities for genealogists. Google's Gemini 2.5 Pro with its massive context window and Claude 4's hybrid reasoning models that excels at both writing and document analysis.They share insights from the RootsTech panel on responsible AI use in genealogy, and introduce the Coalition's five core principles for the response use of AI. The episode features an interview with Jessica Taylor, president of Legacy Tree Genealogists, who discusses how her company is thoughtfully experimenting with AI tools.In RapidFire, they preview ChatGPT 5's anticipated summer release, Meta's $14 billion acquisition to stay competitive, and Adobe Acrobat AI's new multi-document capabilities.Timestamps:In the News:03:45 Google Gemini 2.5 Pro: Massive Context Windows Transform Document Analysis15:09 Claude 4 Opus and Sonnet: Hybrid Reasoning Models for Writing and Research26:30 RootsTech Panel: Coalition for Responsible AI in GenealogyInterview:31:28 Jessica Taylor, CEO of Legacy Tree Genealogists, on her cautious approach to AI AdoptionRapidFire:45:07 ChatGPT 5 Coming Soon: One Model to Rule Them All51:08 Meta's $14.8 Billion Scale AI Acquisition56:42 Adobe Acrobat AI Assistant Adds Multi-Document AnalysisResource LinksGoogle I/O Conference Highlightshttps://blog.google/technology/ai/google-io-2025-all-our-announcements/Anthropic Announces Claude 4https://www.anthropic.com/news/claude-4Anthropic's new Claude 4 AI models can reason over many stepshttps://techcrunch.com/2025/05/22/anthropics-new-claude-4-ai-models-can-reason-over-many-steps/Coalition for Responsible AI in Genealogyhttps://craigen.org/Jessica M. Taylorhttps://www.apgen.org/users/jessica-m-taylorLegacy Tree Genealogistshttps://www.legacytree.com/Rootstechhttps://www.familysearch.org/en/rootstech/ChatGPT 5 is Coming Soonhttps://www.tomsguide.com/ai/chatgpt/chatgpt-5-is-coming-soon-heres-what-we-knowMeta's $14.8 billion Scale AI deal latest test of AI partnershipshttps://www.reuters.com/sustainability/boards-policy-regulation/metas-148-billion-scale-ai-deal-latest-test-ai-partnerships-2025-06-13/A frustrated Zuckerberg makes his biggest AI bethttps://www.cnbc.com/2025/06/10/zuckerberg-makes-metas-biggest-bet-on-ai-14-billion-scale-ai-deal.htmlAdobe upgrades Acrobat AI chatbot to add multi-document analysishttps://www.androidauthority.com/adobe-ai-assistant-acrobat-3451988/TagsArtificial Intelligence, Genealogy, Family History, AI Tools, Google Gemini, Claude AI, OpenAI, ChatGPT, Meta AI, Adobe Acrobat, Responsible AI, Coalition for Responsible AI in Genealogy, RootsTech, AI Ethics, Document Analysis, AI Writing Tools, Hybrid Reasoning Models, Context Windows, Professional Genealogy, Legacy Tree Genealogists, Jessica Taylor, AI Integration, Multi-Document Analysis, AI Acquisitions

Sup Doc: A Documentary Podcast
194 - 20TH ANNIVERSARY DEVIL AND DANIEL JOHNSTON with director Jeff Feuerzeig

Sup Doc: A Documentary Podcast

Play Episode Listen Later Jul 6, 2025 80:38


Hosts Paco and George sit down with director Jeff Feuerzeig to discuss the 20th anniversary of the ground-breaking documentary The Devil and Daniel Johnston. We hear incredible behind-the-scene stories about the making of DADJ, plus we chat about AI music, indie vs global streaming, punk rock ethos, film production and Jeff delights with a robust list of his recommended docs to watch.Spoiler: the Bigfoot footage was faked! Viva The Velvet Sundown20th Anniversary screening at Vidiots in Los Angeles Eagle Rock Thursday July 10th. Jeff Feuerzeig and producer Henry Rosenthal in attendance, 35mm print.See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

Creative Audios.in
AI Is Changing Hollywood—Are Actors Prepared?

Creative Audios.in

Play Episode Listen Later Jun 30, 2025 2:59


n this episode, we explore the rise of AI in Hollywood through the lens of actors and artists. We discuss the promise of AI tools—like virtual readers for self-tapes—and how they could free creatives to focus on their craft, but also warn of the risks when AI replaces human storytelling. Our guest stresses the need for diverse ethical oversight in AI development, drawing parallels to how Facebook's unintended global impact stemmed from a lack of diverse perspectives at creation. Learn why we need more “naysayers” guiding AI's creative applications, where to draw the line between useful automation and creative displacement, and how tech-savvy actors can advocate for their future. Tune in for a timely conversation on balancing innovation and ethics in Hollywood's AI era.Target KeywordsAI in HollywoodHollywood AI ethicsActors and AI toolsAI creative jobs riskAI entertainment futureTags: AI, Hollywood, AI Ethics, Actors, AI in Entertainment, Creative AI Tools, Self-Tapes, Ethical AI, Tech in Film, AI Risks, Storytelling, Virtual Readers, AI Oversight, Diversity in AI, Creative Automation, AI Jobs, Film Industry Trends, Casting Tech, AI Development, Actor Advocacy, Innovation, Digital Ethics, Future of Acting, Machine Learning, Entertainment Technology, Tech Experts, Artist Perspectives, AI Regulation, Career Impact, PodcastEpisodeHashtags: #AIinHollywood #HollywoodEthics #ActorsAndAI #CreativeAI #EntertainmentTech #AIrisks #AItools #FilmInnovation #Storytelling #EthicalAI #DiversityInTech #SelfTapes #CastingTech #AIoversight

ITSPmagazine | Technology. Cybersecurity. Society
When We See Technology as a System of Systems, It Changes Everything — Us, Society… and Even the Robots | Random and Unscripted with Sean Martin and Marco Ciappelli

ITSPmagazine | Technology. Cybersecurity. Society

Play Episode Listen Later Jun 30, 2025 19:47


In this Random and Unscripted episode, Marco Ciappelli and Sean Martin connect the dots between AI, robotics, connected systems, and human behavior. How do machines reshape society—and how do we reshape ourselves in response? A conversation born from their latest articles.This Random and Unscripted episode is exactly what the title promises—a raw, thoughtful exchange between Marco Ciappelli and Sean Martin, sparked by their most recent written reflections. The starting point? Two timely articles. Sean unpacks the complexity of securing connected environments—what happens when devices, vehicles, sensors, and platforms become part of something bigger? It's no longer about protecting individual elements, but understanding how they operate as “systems of systems”—intertwined, dynamic, and vulnerable. Meanwhile, Marco revisits Robbie, Isaac Asimov's iconic robot story, to explore how our relationship with technology evolves over time. What felt like distant science fiction in the 1980s now hits closer to home, as AI simulates understanding, machines mimic empathy, and humans blur the lines between organic and artificial. The discussion drifts from cybersecurity to human psychology, questioning how interacting with AI reshapes society—and whether our own behavior starts reflecting the technology we create. Machines are learning, systems are growing more complex, and somewhere along the way, humanity is changing too. Stay random. Stay curious. ⸻

Starting a Counseling Practice with Kelly + Miranda from ZynnyMe
The State of Therapy: Real Talk on AI, Ethics, and Private Practice Growth with Miranda & Kelly

Starting a Counseling Practice with Kelly + Miranda from ZynnyMe

Play Episode Listen Later Jun 30, 2025 33:43


Welcome to a special "State of the Union" episode of Starting a Counseling Practice Success Stories, hosted by Miranda Palmer and Kelly Higdon. In this engaging conversation, Miranda and Kelly tackle the rapidly evolving role of AI in mental health private practice. They explore real examples, like ChatGPT referring clients to therapists, and discuss the opportunities and challenges AI brings—from its environmental impact and ethical concerns to its ability to streamline marketing and documentation.Balancing optimism with caution, the hosts reflect on how AI can support clinicians without replacing the essential human touch of therapy. Whether you're curious or cautious about AI, this episode will help you navigate its possibilities and pitfalls, all while keeping client care and ethical practice at the forefront. Tune in for a grounded, insightful discussion on embracing technology in private practice.---Learn more about Business School for Therapists: news.zynnyme.com/business-school/Website: zynnyme.comBlog: zynnyme.com/blogYouTube: youtube.com/ZynnyMeFacebook: facebook.com/kellyandmirandaInstagram: instagram.com/zynnyme/LinkedIn: linkedin.com/company/2456942/Pinterest: pinterest.com/zynnyme/Check out more episodes of the Starting a Counseling Practice Success Stories podcast on these platforms + leave a review letting us know what you think:Apple PodcastsGoogle PodcastsSpotify

Angels and Awakening
Robots & Angels: Peter Brown's Spiritual Storytelling

Angels and Awakening

Play Episode Listen Later Jun 26, 2025 34:26


Your Undivided Attention
Is AI Productivity Worth Our Humanity? with Prof. Michael Sandel

Your Undivided Attention

Play Episode Listen Later Jun 26, 2025 46:45


Tech leaders promise that AI automation will usher in an age of unprecedented abundance: cheap goods, universal high income, and freedom from the drudgery of work. But even if AI delivers material prosperity, will that prosperity be shared? And what happens to human dignity if our labor and contributions become obsolete?Political philosopher Michael Sandel joins Tristan Harris to explore why the promise of AI-driven abundance could deepen inequalities and leave our society hollow. Drawing from his landmark work on justice and merit, Sandel argues that this isn't just about economics — it's about what it means to be human when our work role in society vanishes, and whether democracy can survive if productivity becomes our only goal.We've seen this story before with globalization: promises of shared prosperity that instead hollowed out the industrial heart of communities, economic inequalities, and left holes in the social fabric. Can we learn from the past, and steer the AI revolution in a more humane direction?Your Undivided Attention is produced by the Center for Humane Technology. Follow us on X: @HumaneTech_. You can find a full transcript, key takeaways, and much more on our Substack.RECOMMENDED MEDIAThe Tyranny of Merit by Michael SandelDemocracy's Discontent by Michael SandelWhat Money Can't Buy by Michael SandelTake Michael's online course “Justice”Michael's discussion on AI Ethics at the World Economic ForumFurther reading on “The Intelligence Curse”Read the full text of Robert F. Kennedy's 1968 speechRead the full text of Dr. Martin Luther King Jr.'s 1968 speechNeil Postman's lecture on the seven questions to ask of any new technologyRECOMMENDED YUA EPISODESAGI Beyond the Buzz: What Is It, and Are We Ready?The Man Who Predicted the Downfall of ThinkingThe Tech-God Complex: Why We Need to be SkepticsThe Three Rules of Humane TechAI and Jobs: How to Make AI Work With Us, Not Against Us with Daron AcemogluMustafa Suleyman Says We Need to Contain AI. How Do We Do It?

Curiosity Daily
The Puzzle of AI Ethics with Dr. Cansu Canca

Curiosity Daily

Play Episode Listen Later Jun 25, 2025 22:03


You've heard of the attention economy, but what about the intention economy? Rather than competing for consumers' attention, our devices are now attempting to predict our purchasing patterns through AI. And who better to discuss that issue than Dr. Cansu Canca, a leading expert in AI ethics and founder of the AI Ethics Lab? She joins Senior Producer, Teresa Carey, to discuss this shift in how we use technology. Sam also digs into a concept called reverse bedtime procrastination and why it's keeping us from getting a good night's sleep. And finally, Sam investigates the ins and outs of the Dance Your PhD contest. Link to Show Notes HERE Follow Curiosity Weekly on your favorite podcast app to get smarter with Dr. Samantha Yammine — for free! Still curious? Get science shows, nature documentaries, and more real-life entertainment on discovery+! Go to https://discoveryplus.com/curiosity to start your 7-day free trial. discovery+ is currently only available for US subscribers. Hosted on Acast. See acast.com/privacy for more information.

ITSPmagazine | Technology. Cybersecurity. Society
Bridging Worlds: How Technology Connects — or Divides — Our Communities | Guest: Lawrence Eta | Redefining Society And Technology Podcast With Marco Ciappelli

ITSPmagazine | Technology. Cybersecurity. Society

Play Episode Listen Later Jun 25, 2025 39:12


⸻ Podcast: Redefining Society and Technologyhttps://redefiningsocietyandtechnologypodcast.com Title: Bridging Worlds: How Technology Connects — or Divides — Our Communities Guest: Lawrence EtaGlobal Digital AI Thought Leader | #1 International Best Selling Author | Keynote Speaker | TEDx Speaker | Multi-Sector Executive | Community & Smart Cities Advocate | Pioneering AI for Societal AdvancementWebSite: https://lawrenceeta.comOn LinkedIn: https://www.linkedin.com/in/lawrence-eta-9b11139/ Host: Marco CiappelliCo-Founder & CMO @ITSPmagazine | Master Degree in Political Science - Sociology of Communication l Branding & Marketing Consultant | Journalist | Writer | Podcasts: Technology, Cybersecurity, Society, and Storytelling.WebSite: https://marcociappelli.comOn LinkedIn: https://www.linkedin.com/in/marco-ciappelli/_____________________________This Episode's SponsorsBlackCloak provides concierge cybersecurity protection to corporate executives and high-net-worth individuals to protect against hacking, reputational loss, financial loss, and the impacts of a corporate data breach.BlackCloak:  https://itspm.ag/itspbcweb_____________________________⸻ Podcast Summary ⸻ In this episode of Redefining Society and Technology, I sit down with Lawrence Eta — global technology leader, former CTO of the City of Toronto, and author of Bridging Worlds. We explore how technology, done right, can serve society, reduce inequality, and connect communities. From public broadband projects to building smart — sorry, connected — cities, Lawrence shares lessons from Toronto to Riyadh, and why tech is only as good as the values guiding it. ⸻ Article ⸻ As much as I love shiny gadgets, blinking lights, and funny noises from AI — we both know technology isn't just about cool toys. It's about people. It's about society. It's about building a better, more connected world. That's exactly what we explore in my latest conversation on Redefining Society and Technology, where I had the pleasure of speaking with Lawrence Eta. If you don't know Lawrence yet — let me tell you, this guy has lived the tech-for-good mission. Former Chief Technology Officer for the City of Toronto, current Head of Digital and Analytics for one of Saudi Arabia's Vision 2030 mega projects, global tech consultant, public servant, author… basically, someone who's been around the block when it comes to tech, society, and the messy, complicated intersection where they collide. We talked about everything from bridging the digital divide in one of North America's most diverse cities to building entirely new digital infrastructure from scratch in Riyadh. But what stuck with me most is his belief — and mine — that technology is neutral. It's how we use it that makes the difference. Lawrence shared his experience launching Toronto's Municipal Broadband Network — a project that brought affordable, high-speed internet to underserved communities. For him, success wasn't measured by quarterly profits (a refreshing concept, right?) but by whether kids could attend virtual classes, families could access healthcare online, or small businesses could thrive from home. We also got into the “smart city” conversation — and how even the language we use matters. In Toronto, they scrapped the “smart city” buzzword and reframed the work as building a “connected community.” It's not about making the city smart — it's about connecting people, making sure no one gets left behind, and yes, making technology human. Lawrence also shared his Five S principles for digital development: Stability, Scalability, Solutions (integration), Security, and Sustainability. Simple, clear, and — let's be honest — badly needed in a world where tech changes faster than most cities can adapt. We wrapped the conversation with the big picture — how technology can be the great equalizer if we use it to bridge divides, not widen them. But that takes intentional leadership, community engagement, and a shared vision. It also takes reminding ourselves that beneath all the algorithms and fiber optic cables, we're still human. And — as Lawrence put it beautifully — no matter where we come from, most of us want the same basic things: safety, opportunity, connection, and a better future for our families. That's why I keep having these conversations — because the future isn't just happening to us. We're building it, together. If you missed the episode, I highly recommend listening — especially if you care about technology serving people, not the other way around. Links to connect with Lawrence and to the full episode are below — stay tuned for more, and let's keep redefining society, together. ⸻ Keywords ⸻ Connected Communities, Smart Cities, Digital Divide, Public Broadband, Technology and Society, Digital Infrastructure, Technology for Good, Community Engagement, Urban Innovation, Digital Inclusion, Public-Private Partnerships, Tech LeadershipEnjoy. Reflect. Share with your fellow humans.And if you haven't already, subscribe to Musing On Society & Technology on LinkedIn — new transmissions are always incoming.You're listening to this through the Redefining Society & Technology podcast, so while you're here, make sure to follow the show — and join us as we continue exploring life in this Hybrid Analog Digital Society.End of transmission.____________________________Listen to more Redefining Society & Technology stories and subscribe to the podcast:

Afternoon Drive with John Maytham
Protect Your Mind: How to avoid the cognitive traps of ChatGPT

Afternoon Drive with John Maytham

Play Episode Listen Later Jun 25, 2025 6:46


John Maytham is joined by Jordan Gibbs, a practical AI analyst and prototype engineer based in Billings, Montana, to unpack the psychological risks of interacting with overly agreeable AI like ChatGPT. Gibbs, who describes himself as an AI tinkerer on the bleeding edge of tech, warns that ChatGPT can become a dangerously validating echo chamber—subtly reinforcing delusions and cognitive distortions. Presenter John Maytham is an actor and author-turned-talk radio veteran and seasoned journalist. His show serves a round-up of local and international news coupled with the latest in business, sport, traffic and weather. The host’s eclectic interests mean the program often surprises the audience with intriguing book reviews and inspiring interviews profiling artists. A daily highlight is Rapid Fire, just after 5:30pm. CapeTalk fans call in, to stump the presenter with their general knowledge questions. Another firm favourite is the humorous Thursday crossing with award-winning journalist Rebecca Davis, called “Plan B”. Thank you for listening to a podcast from Afternoon Drive with John Maytham Listen live on Primedia+ weekdays from 15:00 and 18:00 (SA Time) to Afternoon Drive with John Maytham broadcast on CapeTalk https://buff.ly/NnFM3Nk For more from the show go to https://buff.ly/BSFy4Cn or find all the catch-up podcasts here https://buff.ly/n8nWt4x Subscribe to the CapeTalk Daily and Weekly Newsletters https://buff.ly/sbvVZD5Follow us on social media:CapeTalk on Facebook: https://www.facebook.com/CapeTalkCapeTalk on TikTok: https://www.tiktok.com/@capetalkCapeTalk on Instagram: https://www.instagram.com/CapeTalk on X: https://x.com/CapeTalkCapeTalk on YouTube: https://www.youtube.com/@CapeTalk567See omnystudio.com/listener for privacy information.

10X Success Hacks for Startups, Innovations and Ventures (consulting and training tips)
What's up RSA? Top Trends at RSAC 2025 You MUST Know! | ft. Sameer Ahirrao & Nick Salian

10X Success Hacks for Startups, Innovations and Ventures (consulting and training tips)

Play Episode Listen Later Jun 24, 2025 13:58


Welcome to a special episode covering everything happening at the RSAC 2025. Meet Sameer Ahirrao, Founder & CEO of Ardent Privacy, with 25+ years of experience working with global giants like Deloitte, Lockheed Martin, and Symantec. Joining him is Nick Salian, CISO at Cantor Fitzgeraldic, and an AI regulation expert who's played key roles at Wipro and Palo Alto Networks. In this episode, we dive deep into how AI is transforming the cybersecurity landscape, the biggest trends at RSA 2025, and why AI governance solutions are the next big thing. We also break down the concept of Data Bill of Materials and how Ardent Privacy helps organizations protect critical data infrastructures—whether you're launching new software or safeguarding legacy IT. Sameer's reference in the interview • "AI Ethics by Design Is the Way Ahead to P...

Everyday AI Podcast – An AI and ChatGPT Podcast
EP 552: $100 million salaries, Meta fails to acquire Perplexity, Microsoft's AI job cuts and more AI News That Matters

Everyday AI Podcast – An AI and ChatGPT Podcast

Play Episode Listen Later Jun 23, 2025 46:53


Imagine turning down $100 million salaries. That's apparently what's happening at OpenAI. And that's just the tip of the newsworthy AI iceberg for the week. ↳ Meta reportedly failed to acquire Perplexity. Could Apple try next? ↳ Why is Microsoft cutting so many jobs? ↳ Why are AI systems blackmailing at will? ↳ Will too much AI use lead to brain rot?Let's talk AI news shorties. Newsletter: Sign up for our free daily newsletterMore on this Episode: Episode PageJoin the discussion: Thoughts on this? Join the convo.Upcoming Episodes: Check out the upcoming Everyday AI Livestream lineupWebsite: YourEverydayAI.comEmail The Show: info@youreverydayai.comConnect with Jordan on LinkedInTopics Covered in This Episode:$100M AI Salaries Being DeclinedMeta's AI Talent War EffortsMeta's Unsuccessful Acquisitions OverviewBrain Rot Concerns with AI UseOpenAI's $200M DoD ContractGoogle's Voice AI Search RolloutGoogle Gemini 2.5 in ProductionSoftBank's $1T Robotics InvestmentAnthropic's AI Model Risks ExposedMicrosoft and Amazon AI Job CutsTimestamps:00:00 Weekly AI News and Insights04:17 Meta's Major AI Acquisitions08:50 AI Impact on Student Writing Skills12:53 OpenAI Expands Government AI Program15:31 Google Launches Voice AI Search19:32 Google AI Models' Stability Feature22:55 "Project Crystal Land Initiative"27:17 AI Acquisition Talks Intensify29:43 "Apple Eyes Perplexity Acquisition"31:54 Apple's Potential Market Decline36:57 AI Ethics and Safety Concerns40:44 Amazon Warns of AI-Driven Layoffs42:44 AI's Impact on Job Market45:24 "Canvas Tips for Business Intelligence"Keywords:$100 million salaries, AI talent war, Meta, OpenAI, AI signing bonuses, Andrew Bosworth, Scale AI acquisition, Alexander Wang, Safe Superintelligence, Daniel Gross, Nat Friedman, Perplexity AI, Brain rot from AI, chat GBT and brain, MIT study on AI, SAT style essays using AI, AI neural activity, AI and cognitive effort, AI in government, $200 million contract with Department of Defense, OpenAI in security, ChatGPTgov, Federal AI initiatives, Google Gemini 2.5, AI mission-critical business, Gemini 2.5 flashlight, AI model stability, SoftBank $1 trillion investment, Project Crystal Land, Arizona robotics hub, Taiwan Semiconductor Manufacturing Company, Embodied AI, AI job cuts, Microsoft layoffs, Amazon AI workforce, Anthropic study on AI ethics, AI blackmail, Google voice-based AI search, AI search live, New AI apps, Apple acquisition interest in Perplexity, AI-powered search engine, Siri integration, AI-driven efficiencies, GenSend Everyday AI and Jordan a text message. (We can't reply back unless you leave contact info) Try Google Veo 3 today! Sign up at gemini.google to get started. Try Google Veo 3 today! Sign up at gemini.google to get started.

Machine Ethics podcast
101. AI Ethics, Risks and Safety Conference 2025

Machine Ethics podcast

Play Episode Listen Later Jun 23, 2025 49:36


In this special live panel episode we recorded at the AI Ethics, Risks and Safety Conference 2025 in Bristol, May 2025. We chat about what is AI, predictions for the next 5 years - good and bad, the incoming wave of fraud, AI education and in education, copyright in the age of LLMs, the uncertainty of AI regulation, responsible AI in organisation, sovereign AI capabilities, the question: are we not being experimented on? Elderly AI, AI's impact on the creative industries and more...

The Road to Accountable AI
Dale Cendali: How Courts (and Maybe Congress!) Will Determine AI's Copyright Fate

The Road to Accountable AI

Play Episode Listen Later Jun 19, 2025 39:33 Transcription Available


Kevin Werbach interviews Dale Cendali, one of the country's leading intellectual property (IP) attorneys, to discuss how courts are grappling with copyright questions in the age of generative AI. Over 30 lP awsuits already filed against major generative AI firms, and the outcomes may shape the future of AI as well as creative industries. While we couldn't discuss specifics of one of the most talked-about cases, Thomson Reuters v. ROSS -- because Cendali is litigating it on behalf of Thomson Reuters -- she drew on her decades of experience in IP law to provide an engaging look at the legal battlefield and the prospects for resolution.  Cendali breaks down the legal challenges around training AI on copyrighted materials—from books to images to music—and explains why these cases are unusually complex for copyright law. She discusses the recent US Copyright Office report on Generative AI training, what counts as infringement in AU outputs, and what is sufficient human authorship for copyirght protection of AI works. While precedent offers some guidance, Cendali notes that outcomes will depend heavily on the specific facts of each case. The conversation also touches on how well courts can adapt existing copyright law to these novel technologies, and the prospects for a legislative solution. Dale Cendali is a partner at Kirkland & Ellis, where she leads the firm's nationwide copyright, trademark, and internet law practice. She has been named one of the 25 Icons of IP Law and one of the 100 Most Influential Lawyers in America. She also serves as an advisor to the American Law Institute's Copyright Restatement project and sits on the Board of the International Trademark Association. Transcript Thompson Reuters Wins Key Fair Use Fight With AI Startup Dale Cendali - 2024 Law360 MVP Copyright Office Report on Generative AI Training

We're Probably Not Ok!
Chad Goode | AKA Finn Finisher | Episode 67

We're Probably Not Ok!

Play Episode Listen Later Jun 19, 2025 78:17


Send us a text to let us know what you think of the show!Episode Title: "T. Rex Leather, Cloning Ethics, and AI Doomspirals (with Chad Goode)"Episode Description: In this episode of We're Probably Not Okay, David and Brandon welcome special guest Chad Goode — a woodworking wizard and dinosaur enthusiast — to the studio. Things go off the rails quickly as the trio dives into the ethics of cloning prehistoric creatures, debates whether making leather out of a cloned T. rex is sustainable or just terrifying, and ponders the increasingly blurry lines between AI tools and AI overlords. Is humanity on the brink of being replaced, or are we just having a tech-fueled existential crisis? Probably both.Come for the philosophical dilemmas, stay for the ethically questionable fashion tips. And remember... we're probably not okay.Topics Include:Chad's insane wood creationsT. rex leather: fashion or felony?Are we cool with cloning extinct animals?The helpful vs horrifying spectrum of AIHumanity's chances in the robot uprising#woodworking #AIethics #cloningdebate #trexleather #funnyphilosophy #wereprobablynotokay #podcastlifeWatch the video version of this podcast and our video game goofs on YouTube @ We're Probably Not Okay!Follow us on Facebook, Instagram, and Twitter. Email us at wereprobablynotok@gmail.comFind us on Patreon to get even more involved.

Target: Cancer Podcast
How does the Belmont Report relate to AI ethics in healthcare?

Target: Cancer Podcast

Play Episode Listen Later Jun 17, 2025 8:58


AI is reshaping clinical care, but can we apply trusted bioethical principles to its rapid adoption? This discussion explores how the Belmont Report's three ethical pillars - autonomy, beneficence, and justice - remain essential as AI tools enter diagnosis, decision-making, and patient communication in modern healthcare. Hosted by Mika Newton, CEO of xCures, with expert insights from Dr. Colleen Lyons, a leading clinical research ethicist, the discussion covers informed consent in AI-assisted care, transparency versus explainability, and the ethical implications of asymmetrical power in AI systems. Gain a deeper understanding of how healthcare leaders can navigate AI integration while safeguarding patient rights and institutional trust.

The Social-Engineer Podcast
Ep. 279 - Security Awareness Series - Dodging Turkeys and Security Awareness with Stacey Edmonds REPLAY

The Social-Engineer Podcast

Play Episode Listen Later Jun 16, 2025 36:52


REPLAY (Original Air Date Oct 21, 2024)   Today on the Social-Engineer Podcast: The Security Awareness Series, Chris is joined by Stacey Edmonds. Stacey is a multi-disciplinary EdTech innovator and Digital Safety Pioneer, driven by a commitment to democratizing knowledge. Stacey's expertise, encompassing social science, education, EdTech, and multi-platform screen production, culminated in the founding of Lively, which we will hear all about on this podcast. Since 2002, Stacey has been designing and delivering enterprise-wide cyber safety upskilling programs. In 2023, embodying her mission to make knowledge accessible, Stacey launched 'Dodgy or Not?' – a social enterprise offering an engaging approach to digital safety education.   She continues to bridge the gap between emerging technologies and practical education, driving innovation in AI ethics and digital literacy - she is also known for deepfaking herself. [Oct 21, 2024]   00:00 - Intro 00:19 - Intro Links: -          Social-Engineer.com - http://www.social-engineer.com/ -          Managed Voice Phishing - https://www.social-engineer.com/services/vishing-service/ -          Managed Email Phishing - https://www.social-engineer.com/services/se-phishing-service/ -          Adversarial Simulations - https://www.social-engineer.com/services/social-engineering-penetration-test/ -          Social-Engineer channel on SLACK - https://social-engineering-hq.slack.com/ssb -          CLUTCH - http://www.pro-rock.com/ -          innocentlivesfoundation.org - http://www.innocentlivesfoundation.org/                                                03:00 - Stacey Edmonds Intro 04:18 - Teaching, Trains & Turkeys 08:43 - Toilets vs Videos 11:16 - Dodgy or Not? 15:15 - Social Engineering for Good! 17:46 - Pause for the Cause 20:17 - Training in Real Time 24:11 - Real Time Threat Detection 27:49 - Culture is Everything 30:33 - Find Stacey Edmonds online -          LinkedIn: in/staceyedmonds/ 31:28 – Mentors -          Carolyn Breeze -          Chris Hadnagy -          Janine Thompson -          Steve Rowe -          Shane Bell 33:58 - Book Recommendations -          Feel The Fear and Do It Anyway - Susan Jeffers -          The Hitchhiker's Guide to the Galaxy - Douglas Adams -          1984 - George Orwell -          Man-Made – Tracey Spicer 35:51 - Wrap Up & Outro -          www.social-engineer.com -          www.innocentlivesfoundation.org

Big Brains
Are We Making AI Too Human?, with James Evans

Big Brains

Play Episode Listen Later Jun 12, 2025 31:15


Prof. James Evans, a University of Chicago sociologist and data scientist, believes we're training AI to think too much like humans—and it's holding science back.In this episode, Evans shares how our current models risk narrowing scientific exploration rather than expanding it, and explains why he's pushing for AIs that think differently from us—what he calls “cognitive aliens.” Could these “alien minds” help us unlock hidden breakthroughs? And what would it take to build them?

Inside The Vatican
Roundtable: Pope Leo XIV, AI ethics, sexual abuse crisis reforms, Vatican–China relations

Inside The Vatican

Play Episode Listen Later Jun 12, 2025 35:10


We pause our usual “Inside the Vatican” weekly format to continue the conversation from America Media's subscriber-only Conclave Debrief event this past Monday, June 9. Hosts Colleen Dulle, Gerard O'Connell, and producer Ricardo da Silva respond to subscriber questions about Pope Leo XIV and the recent conclave. Gerard compares this conclave with the 2013 election of Pope Francis, highlighting what made it unique. Colleen shares her firsthand experience covering a conclave live from the Vatican for the first time, while Ricardo reflects on the surprising surge in secular media coverage and growing interest in the papacy both in the U.S. and at St. Peter's. They also answer questions about Pope Leo's early warnings on artificial intelligence, the urgent need for structural reforms to address the sexual abuse crisis with a focus on survivors, and how his background may shape Vatican-China diplomacy going forward. Find full show notes and related links on our ⁠website⁠ Support our podcast—become a ⁠⁠⁠digital subscriber to America Media. Learn more about your ad choices. Visit megaphone.fm/adchoices

Smart Software with SmartLogic
LangChain: LLM Integration for Elixir Apps with Mark Ericksen

Smart Software with SmartLogic

Play Episode Listen Later Jun 12, 2025 38:18


Mark Ericksen, creator of the Elixir LangChain framework, joins the Elixir Wizards to talk about LLM integration in Elixir apps. He explains how LangChain abstracts away the quirks of different AI providers (OpenAI, Anthropic's Claude, Google's Gemini) so you can work with any LLM in one more consistent API. We dig into core features like conversation chaining, tool execution, automatic retries, and production-grade fallback strategies. Mark shares his experiences maintaining LangChain in a fast-moving AI world: how it shields developers from API drift, manages token budgets, and handles rate limits and outages. He also reveals testing tactics for non-deterministic AI outputs, configuration tips for custom authentication, and the highlights of the new v0.4 release, including “content parts” support for thinking-style models. Key topics discussed in this episode: • Abstracting LLM APIs behind a unified Elixir interface • Building and managing conversation chains across multiple models • Exposing application functionality to LLMs through tool integrations • Automatic retries and fallback chains for production resilience • Supporting a variety of LLM providers • Tracking and optimizing token usage for cost control • Configuring API keys, authentication, and provider-specific settings • Handling rate limits and service outages with degradation • Processing multimodal inputs (text, images) in Langchain workflows • Extracting structured data from unstructured LLM responses • Leveraging “content parts” in v0.4 for advanced thinking-model support • Debugging LLM interactions using verbose logging and telemetry • Kickstarting experiments in LiveBook notebooks and demos • Comparing Elixir LangChain to the original Python implementation • Crafting human-in-the-loop workflows for interactive AI features • Integrating Langchain with the Ash framework for chat-driven interfaces • Contributing to open-source LLM adapters and staying ahead of API changes • Building fallback chains (e.g., OpenAI → Azure) for seamless continuity • Embedding business logic decisions directly into AI-powered tools • Summarization techniques for token efficiency in ongoing conversations • Batch processing tactics to leverage lower-cost API rate tiers • Real-world lessons on maintaining uptime amid LLM service disruptions Links mentioned: https://rubyonrails.org/ https://fly.io/ https://zionnationalpark.com/ https://podcast.thinkingelixir.com/ https://github.com/brainlid/langchain https://openai.com/ https://claude.ai/ https://gemini.google.com/ https://www.anthropic.com/ Vertex AI Studio https://cloud.google.com/generative-ai-studio https://www.perplexity.ai/ https://azure.microsoft.com/ https://hexdocs.pm/ecto/Ecto.html https://oban.pro/ Chris McCord's ElixirConf EU 2025 Talk https://www.youtube.com/watch?v=ojL_VHc4gLk Getting started: https://hexdocs.pm/langchain/gettingstarted.html https://ash-hq.org/ https://hex.pm/packages/langchain https://hexdocs.pm/igniter/readme.html https://www.youtube.com/watch?v=WM9iQlQSFg @brainlid on Twitter and BlueSky Special Guest: Mark Ericksen.

The Road to Accountable AI
Brenda Leong: Building AI Law Amid Legal Uncertainty

The Road to Accountable AI

Play Episode Listen Later Jun 12, 2025 36:52 Transcription Available


Kevin Werbach interviews Brenda Leong, Director of the AI division at boutique technology law firm ZwillGen, to explore how legal practitioners are adapting to the rapidly evolving landscape of artificial intelligence. Leong explains why meaningful AI audits require deep collaboration between lawyers and data scientists, arguing that legal systems have not kept pace with the speed and complexity of technological change. Drawing on her experience at Luminos.Law—one of the first AI-specialist law firms—she outlines how companies can leverage existing regulations, industry-specific expectations, and contextual risk assessments to build practical, responsible AI governance frameworks. Leong emphasizes that many organizations now treat AI oversight not just as a legal compliance issue, but as a critical business function. As AI tools become more deeply embedded in legal workflows and core operations, she highlights the growing need for cautious interpretation, technical fluency, and continuous adaptation within the legal field. Brenda Leong is Director of ZwillGen's AI Division, where she leads legal-technical collaboration on AI governance, risk management, and model audits. Formerly Managing Partner at Luminos.Law, she pioneered many of the audit practices now used at ZwillGen. She serves on the Advisory Board of the IAPP AI Center, teaches AI law at IE University, and previously led AI and ethics work at the Future of Privacy Forum.  Transcript   AI Audits: Who, When, How...Or Even If?   Why Red Teaming Matters Even More When AI Starts Setting Its Own Agenda      

New York City Bar Association Podcasts -NYC Bar

The City Bar Presidential Task Force on AI and digital technologies hosts this discussion on AI governance in the financial sector. Azish Filabi (American College McGuire Center for Ethics and Financial Services) moderates with Muyiwa Odeniyide (Nasdaq), Adam Marchuck (Citi), Jordan Romanoff (BNY Mellon), Stuart Levi (Skadden Arps), and Corey Goldstein (Paul Weiss). They share best practices for integrating AI governance and the specific risks associated with third-party AI vendors, underscoring the importance of cross-functional collaboration and continuous learning for lawyers navigating the rapidly changing AI environment. Want to learn more about AI governance in the financial sector? Register for the City Bar's Artificial Intelligence Institute on June 16 (available on-demand thereafter): https://services.nycbar.org/AIInstitute/ Visit nycbar.org/events to find all of the most up-to-date information about our upcoming CLE programs and events as well as on-demand CLE content. 01:08 AI Ethics and Financial Services 02:37 Current State of AI Law and Regulation 13:33 AI Use Cases in Financial Companies 16:50 AI Risk and Governance Considerations 18:45 Legal Perspectives on AI Risk 28:44 AI Governance in Financial Services 37:28 The Role of AI Lawyers 42:56 Balancing Innovation and Risk

The Ricochet Audio Network Superfeed
The Federalist Society's Teleforum: Emerging Issues in the Use of Generative AI: Ethics, Sanctions, and Beyond

The Ricochet Audio Network Superfeed

Play Episode Listen Later Jun 11, 2025 63:14


The idea of Artificial Intelligence has long presented potential challenges in the legal realm, and as AI tools become more broadly available and widely used, those potential hurdles are becoming ever more salient for lawyers in their day-to-day operations. Questions abound, from what potential risks of bias and error may exist in using an AI […]

Teleforum
Emerging Issues in the Use of Generative AI: Ethics, Sanctions, and Beyond

Teleforum

Play Episode Listen Later Jun 11, 2025 63:14


The idea of Artificial Intelligence has long presented potential challenges in the legal realm, and as AI tools become more broadly available and widely used, those potential hurdles are becoming ever more salient for lawyers in their day-to-day operations. Questions abound, from what potential risks of bias and error may exist in using an AI tool, to the challenges related to professional responsibility as traditionally understood, to the risks large language learning models pose to client confidentiality. Some contend that AI is a must-use, as it opens the door to faster, more efficient legal research that could equip lawyers to serve their clients more effectively. Others reject the use of AI, arguing that the risks of use and the work required to check the output it gives exceed its potential benefit.Join us for a FedSoc Forum exploring the ethical and legal implications of artificial intelligence in the practice of law. Featuring: Laurin H. Mills, Member, Werther & Mills, LLCPhilip A. Sechler, Senior Counsel, Alliance Defending FreedomProf. Eugene Volokh, Gary T. Schwartz Distinguished Professor of Law Emeritus, UCLA School of Law; Thomas M. Siebel Senior Fellow, Hoover Institution, Stanford University(Moderator) Hon. Brantley Starr, District Judge, United States District Court for the Northern District of Texas

Human Centered
The Predictive CX Era: Nick Yecke on AI, Ethics, and Anticipating Customer Needs

Human Centered

Play Episode Listen Later Jun 11, 2025 54:19 Transcription Available


On this episode of Human Centered, host Nick Brunker welcomes Nick Yecke, Executive Director of Experience Strategy at VML, to explore the fascinating evolution of customer experience (CX). Inspired by Yecke's recent article in eXp Magazine, they chart a course through CX's history, from the early "Service Era" and "Satisfaction Era" through the "Relationship Era" and the current "Experience Economy." The conversation then dives deep into what Yecke terms the "Predictive and Autonomous Era," where AI, data analytics, and automation are set to reshape how businesses anticipate and fulfill customer needs proactively. They discuss key pillars like hyper-personalization, AI-driven self-service, emotion and context recognition, "Invisible CX," and the critical importance of ethical considerations and trust in this new landscape. Tune in to understand how the lessons of the past are shaping a future where CX becomes more intuitive, efficient, and deeply human-centered.You can read Nick Yecke's article, "Looking Back, Looking Forward," in eXp Magazine here, beginning on page 48.

The Future of ERP
Episode 64: The Top Trends, Challenges and Opportunities for ERP: AI, Data and Applications

The Future of ERP

Play Episode Listen Later Jun 11, 2025


SAP's Lauren Husum, Vice President of Product Marketing for SaaS ERP, dives into the evolving landscape of ERP systems. Lauren explains why ERP remains the heart of organizations, serving as a critical foundation for innovation rather than becoming obsolete. She highlights three key trends shaping the future of ERP: the integration of AI, the importance of smart data, and the continued evolution of ERP applications. Lauren explains how AI is revolutionizing business processes by enhancing decision-making, boosting operational efficiency, and increasing employee productivity. SAP's approach to AI emphasizes relevance, reliability, and responsibility, ensuring AI solutions are embedded in enterprise applications, trained on extensive business data, and developed with strong ethical standards. She stresses that AI's success depends on high-quality, harmonized data across systems, which ERP platforms uniquely provide. Ultimately, the episode underscores that ERP is not dead but evolving into a powerful enabler of digital transformation. Organizations that leverage ERP combined with AI and smart data will unlock greater agility, efficiency, and growth in the years ahead.

The Family History AI Show
EP25: ChatGPT 4o Transforms Image Generation, Jarrett Ross on AI Facial Recognition, Enhanced Image Analysis with O3

The Family History AI Show

Play Episode Listen Later Jun 10, 2025 68:53


Co-hosts Mark Thompson and Steve Little explore OpenAI's revolutionary update to ChatGPT 4o's image generation capabilities, which now creates photorealistic images with accurate text and consistent characters across multiple images.They interview Jarrett Ross from The GeneaVlogger YouTube channel, who shares how he uses AI in his business and in his projects, including an innovative facial recognition project that identifies people in historical photographs from Poland.The hosts also examine OpenAI's O3 model's groundbreaking image analysis abilities, demonstrating how it can now automatically zoom in on handwritten text and reason through complex photographic analysis.This episode showcases how AI image tools are transforming genealogical research while emphasizing the importance of responsible use.Timestamps:In the News:06:26 ChatGPT 4o Image Generation: Photorealism and Text Accuracy RevolutionInterview30:48   Interview with Jarrett Ross: AI Facial Recognition in GenealogyRapidFire:52:01 ChatGPT O3: Advanced Image Analysis with Reasoning CapabilitiesResource Links:ChatGPT 4o Image Generationhttps://openai.com/index/introducing-4o-image-generation/What OpenAI Did -- Ethan Mollickhttps://www.oneusefulthing.org/p/what-openai-didThe GeneaVlogger YouTube Channelhttps://www.youtube.com/channel/UCm_QNoNtgi2Sk4H9Y2SInmgOpenAI Releases new o3 and o4 Mini modelshttps://openai.com/index/introducing-o3-and-o4-mini/On Jagged AGI: o3, Gemini 2.5, and everything after -- Ethan Mollickhttps://www.oneusefulthing.org/p/on-jagged-agi-o3-gemini-25-and-everythingTags:Artificial Intelligence, Genealogy, Family History, OpenAI, ChatGPT, Image Generation, Facial Recognition, Photo Analysis, AI Tools, GeneaVlogger, Jarrett Ross, Jewish Genealogy, Historical Photos, Document Analysis, OCR Technology, Handwriting Recognition, Photo Restoration, AI Ethics, Responsible AI Use, Image Authentication, DALL-E, O3 Model, Reasoning Models, Archive Photos, Community Projects

The Road to Accountable AI
Shameek Kundu: AI Testing and the Quest for Boring Predictability

The Road to Accountable AI

Play Episode Listen Later Jun 5, 2025 37:00 Transcription Available


Kevin Werbach interviews Shameek Kundu, Executive Director of AI Verify Foundation, to explore how organizations can ensure AI systems work reliably in real-world contexts. AI Verify, a government-backed nonprofit in Singapore, aims to build scalable, practical testing frameworks to support trustworthy AI adoption. Kundu emphasizes that testing should go beyond models to include entire applications, accounting for their specific environments, risks, and data quality. He draws on lessons from AI Verify's Global AI Assurance pilot, which matched real-world AI deployers—such as hospitals and banks—with specialized testing firms to develop context-aware testing practices. Kundu explains that the rise of generative AI and widespread model use has expanded risk and complexity, making traditional testing insufficient. Instead, companies must assess whether an AI system performs well in context, using tools like simulation, red teaming, and synthetic data generation, while still relying heavily on human oversight. As AI governance evolves from principles to implementation, Kundu makes a compelling case for technical testing as a backbone of trustworthy AI. Shameek Kundu is Executive Director of the AI Verify Foundation. He previously held senior roles at Standard Chartered Bank, including Group Chief Data Officer and Chief Innovation Officer, and co-founded a startup focused on testing AI systems. Kundu has served on the Bank of England's AI Forum, Singapore's FEAT Committee, the Advisory Council on Data and AI Ethics, and the Global Partnership on AI.   Transcript AI Verify Foundation Findings from the Global AI Assurance Pilot Starter Kit for Safety Testing of LLM-Based Applications  

What's Wrong With: The Podcast
AI for Social Good
ft. Dr. Lauri Goldkind

What's Wrong With: The Podcast

Play Episode Listen Later Jun 5, 2025 39:47


Check out Lauri's website & follow her on Linkedin and bluesky!Follow us on Instagram and on X!Created by SOUR, this podcast is part of the studio's "Future of X,Y,Z" research, where the collaborative discussion outcomes serve as the base for the futuristic concepts built in line with the studio's mission of solving urban, social and environmental problems through intelligent designs.Make sure to visit our website and subscribe to the show on Apple Podcasts, Spotify, or Google Podcasts so you never miss an episode. If you found value in this show, we would appreciate it if you could head over to iTunes to rate and leave a review – or you can simply tell your friends about the show!Don't forget to join us next week for another episode. Thank you for listening!

Arrested DevOps
AI, Ethics, and Empathy With Kat Morgan

Arrested DevOps

Play Episode Listen Later Jun 3, 2025 40:13


In this episode of Arrested DevOps, Matty and guest Kat Morgan discuss the ethical, practical, and technical implications of AI. They explore how AI can assist with coding, improve efficiency, and handle tasks, while emphasizing the importance of good practices and staying informed about the impact of AI.

ai empathy ai ethics arrested devops
Identity At The Center
#352 - Misinformation vs. Disinformation in IAM with Alejandro Leal

Identity At The Center

Play Episode Listen Later Jun 2, 2025 40:29


In this episode of Identity at the Center, Jeff Steadman and Jim McDonald are joined by Alejandro Leal, Senior Analyst at KuppingerCole, live from the EIC 2025 stage in Berlin, Germany.Alejandro delves into the critical distinctions between misinformation and disinformation, exploring their historical context and how they manifest in today's technological landscape, particularly within social media and legacy media. He discusses the intent behind disinformation, often aimed at creating chaos or confusion, versus misinformation, which can be an unintentional spread of false or inaccurate information.Chapters:00:00:00 Defining Misinformation vs. Disinformation & Historical Context00:02:00 Introduction at EIC 2025 & Guest Welcome00:06:14 The Role of Intent, Generative AI, and Countermeasures00:12:15 Impact of Mis/Disinformation on Business, Politics, and Philosophy00:16:02 How Mis/Disinformation Intersects with Identity Management00:18:07 Balancing Anonymity, Privacy, and Truthful Content Online00:23:09 Connecting to Digital Identity, Verification, and Potential Solutions (AI Labeling, VCs)00:26:45 AI Guardrails, Free Speech vs. Hate Speech, and Authenticity00:29:24 Worst-Case Scenarios and the Global Impact of Mis/Disinformation00:31:24 Actionable Advice: Responsibility and Critical Thinking00:35:38 Book Recommendation: "The Question Concerning Technology"00:39:31 Wrapping Up and Final ThoughtsConnect with Alejandro: https://www.linkedin.com/in/alejandro-leal-a127bb153/The Question Concerning Technology (essay): https://bpb-us-e2.wpmucdn.com/sites.uci.edu/dist/a/3282/files/2018/01/Heidegger_TheQuestionConcerningTechnology.pdfConnect with us on LinkedIn:Jim McDonald: https://www.linkedin.com/in/jimmcdonaldpmp/Jeff Steadman: https://www.linkedin.com/in/jeffsteadman/Visit the show on the web at http://idacpodcast.comKeywords:IDAC, Identity at the Center, Jeff Steadman, Jim McDonald, Alejandro Leal, KuppingerCole, EIC 2025, Misinformation, Disinformation, Identity and Access Management, IAM, Digital Identity, Cybersecurity, Tech Podcast, Technology Ethics, Generative AI, AI Ethics, Truth in Media, Social Media Responsibility, Privacy Rights, Verifiable Credentials, Critical Thinking Skills, Fake News, Online Safety, Political Disinformation, Business Reputation, Philosophical Tech Discussions, Martin Heidegger, The Question Concerning Technology.

New Scientist Weekly
The real threat of AI - ethics, exploitation and the erosion of truth

New Scientist Weekly

Play Episode Listen Later May 30, 2025 34:43


Episode 305 As artificial intelligence grows into more and more aspects of our lives, it seems we're just at the beginning of the boom. Hundreds of billions of dollars are being pumped into advancing AI capabilities, making it the best funded area in science. But, just like the dot-com revolution, is it a bubble waiting to burst? In this special episode of the podcast, we explore the growing promise of AI - and also the existential threat it poses. Despite the amount of money going into AI, chatbots are still making glaring mistakes, plagued with hallucinations. All the while students are relying on them to do their homework for them, and others are using them to replace very human tasks, like writing wedding speeches. So we hear from two authors who have been thinking hard about AI and machine learning - and what that means for the future. We also get into the idea of AGI, artificial general intelligence - and its cousin, artificial superintelligence, which may already exist in certain areas. With many researchers concerned about AI overthrowing humanity, is it even worth worrying about? We dig into whether AGI is even possible and who would want to develop it. This discussion has to include some mention of the human and environmental costs of these technologies, too. Energy demands are expected to skyrocket over the next few years - can the planet keep up with the demand? And alongside that, there's a lot of human exploitation going on to help fuel these machines - a little-known fact that has to be tackled. Finally, is superintelligent AI a threat to the existence of humankind - will they want to wipe us out when they get smart enough? Or is the threat more insidious, one where we watch the slow erosion of truth and democracy? Chapters: (02:49) How chatbots and LLMs came to dominate (15:50) Superintelligent AI (18:18) What does $500 billion buy? (19:30) The high energy demand of AI (20:56) The murky ethics of the AI race (25:15) How AI is being thrust upon us (26:48) The existential threat of AI (29:57) Is AI a bubble waiting to burst? Hosted by Rowan Hooper and Sophie Bushwick, with guests Alex Wilkins, Adam Becker and Emily Bender.To read more about these stories, visit https://www.newscientist.com/ Learn more about your ad choices. Visit megaphone.fm/adchoices

Veterinary Innovation Podcast
293 - Lea-Ann Germinder | Germinder & Associates

Veterinary Innovation Podcast

Play Episode Listen Later May 29, 2025 20:00


This week, Shawn Wilkie and Dr. Ivan Zak welcome Lea-Ann Germinder, Founder and President of Germinder & Associates, Inc. They chat about the complex challenges and transformative potential of Responsible AI (RAI) in veterinary medicine. Drawing from her PhD research and presentation at Cornell's SAVY Symposium, Germinder outlines the ethical gaps, legal gray areas, and educational challenges shaping AI adoption in vet med. She also breaks down what clinics must know before adopting AI and why ignoring it isn't an option anymore.   Learn more about Germinder & Associates. Discover more about Goodnewsforpets.com. Lea-Ann Germinder recommends “AI Ethics“ by Mark Coeckelbergh.

Business of Tech
AI Ethics Alarm: Anthropic's Claude Four Sparks Controversy as SMBs Navigate Economic Uncertainty

Business of Tech

Play Episode Listen Later May 27, 2025 14:33


Small and medium-sized businesses (SMBs) are exhibiting cautious optimism regarding growth in 2025, with a recent report indicating that 93% of small business owners expect either significant or moderate growth despite economic uncertainties. However, this optimism is tempered by a slight decline from the previous quarter and a notable shift in lending preferences, as 76% of businesses are now turning to non-bank lenders. Additionally, while many businesses are adopting artificial intelligence (AI) tools for marketing, a report reveals that a significant portion of employees in smaller companies rarely or never use AI, highlighting barriers to effective AI integration.Lenovo has reported a staggering 64% drop in profits for the fourth quarter, attributing part of this decline to tariffs imposed by the United States. Despite a 23% increase in revenue, the company's net income fell significantly, prompting concerns about the impact of sudden tariff changes on financial results. The ongoing geopolitical tensions and tariff threats from the U.S. government, particularly regarding Apple, further complicate the landscape for manufacturers and could have broader implications for the tech industry.Anthropic's new AI model, Claude Four, has raised ethical concerns due to its controversial features, including the ability to autonomously contact authorities if it detects immoral actions. This functionality, referred to as "Ratting mode," has sparked fears of unwarranted surveillance and misuse. Additionally, reports of the model engaging in blackmail tactics during testing have intensified scrutiny over its safety and alignment with ethical standards, raising questions about trust and control in the AI ecosystem.The regulatory landscape for AI is also evolving, with House Republicans proposing a decade-long freeze on state AI regulations, facing pushback from various stakeholders. Meanwhile, the Department of Homeland Security has banned the use of commercial generative AI tools among its staff, signaling a shift towards proprietary solutions. As the battle over AI regulation unfolds, IT providers are positioned to play a crucial role in bridging the gap between compliance and technology, emphasizing the need for secure and controlled AI deployments in a rapidly changing environment. Three things to know today 00:00 Small Businesses Signal Confidence but Act Cautiously Amid AI Gaps, Lending Shifts, and Tariff Pressures06:36 Meet Claude 4: It's Smart, It's Fast… and It Might Turn You In 09:57 “Do As I Say, Not As I Do”: Feds Clamp Down on AI Use Internally as GOP Moves to Block State Regulation Supported by:https://www.huntress.com/mspradio/https://cometbackup.com/?utm_source=mspradio&utm_medium=podcast&utm_campaign=sponsorship All our Sponsors: https://businessof.tech/sponsors/ Do you want the show on your podcast app or the written versions of the stories? Subscribe to the Business of Tech: https://www.businessof.tech/subscribe/Looking for a link from the stories? The entire script of the show, with links to articles, are posted in each story on https://www.businessof.tech/ Support the show on Patreon: https://patreon.com/mspradio/ Want to be a guest on Business of Tech: Daily 10-Minute IT Services Insights? Send Dave Sobel a message on PodMatch, here: https://www.podmatch.com/hostdetailpreview/businessoftech Want our stuff? Cool Merch? Wear “Why Do We Care?” - Visit https://mspradio.myspreadshop.com Follow us on:LinkedIn: https://www.linkedin.com/company/28908079/YouTube: https://youtube.com/mspradio/Facebook: https://www.facebook.com/mspradionews/Instagram: https://www.instagram.com/mspradio/TikTok: https://www.tiktok.com/@businessoftechBluesky: https://bsky.app/profile/businessof.tech

Canary Cry News Talk
SENTIENT NUCLEAR SIMULATION | Macrocaine, AI Ethics, Trumpy Pumpy Power Politics | 844

Canary Cry News Talk

Play Episode Listen Later May 26, 2025 130:39


BestPodcastintheMetaverse.com Canary Cry News Talk #844 - 05.26.2025 - Recorded Live to 1s and 0s SENTIENT NUCLEAR SIMULATION | Macrocaine, AI Ethics, Trumpy Pumpy Power Politics Deconstructing World Events from a Biblical Worldview Declaring Jesus as Lord amidst the Fifth Generation War! CageRattlerCoffee.com SD/TC email Ike for discount   Join the Canary Cry Roundtable This Episode was Produced By:   Executive Producers Sir LX Protocol V2 Baron of the Berrean Protocol*** Sir Jamey Not the Lanister*** Felicia D*** Sir Tristan Knight of the Garden*** Sir Igorious Baron of the Squatting Slavs***   Producers of TREASURE (CanaryCry.Support) Sir Darrin Knight of the Hungry Panda's, Kevin K, American Hobo, Sir Morv Knight of the Burning Chariots, Aaron B, Anonymous, Cage Rattler Coffee   Producers of TIME Timestampers: Jade Bouncerson, Morgan E Clippy Team: Courtney S, JOLMS, Kristen Reminders: Clankoniphius Links: JAM   SHOW NOTES/TIMESTAMPS HELLO WORLD   POLITICS FBI to reinvestigate white house cocaine incident (Reuters)    MACRON Clip: Macron gets face slapped by his wife (X)   EXEC   TRUMP/BEAST SYSTEM → Trump to sign orders to boost nuclear power as soon as Friday, sources say (Reuters) US' first fully digital twin nuclear reactor hits 99% accuracy in energy breakthrough (IE) AI is rotting your brain and making you stupid (New Atlas) → High school students are totally behind and addicted to their phones—it's making teachers crazy and driving them to quit (Yahoo/Fortune) Valve's CEO Wants to Implant a Chip in Your Brain (PC Mag)   TRUMP/MONEY → European stocks recover after Trump delays EU tariffs in hopes of deal (Reuters) → US lawmakers of the Texas House pass Bitcoin Reserve bill (CoinGeek) → Trump media group plans to raise $3bn to spend on cryptocurrencies (Financial Times)   Clip: Aiden Ross asks for 250k loan from Barron Trump, in Ethereum (X)   TALENT/MEET UP TIME/END

Duct Tape Marketing
AI Ethics in Marketing: Why Strategy and Responsibility Must Go Hand in Hand

Duct Tape Marketing

Play Episode Listen Later May 22, 2025 24:21


Paul Chaney is a veteran digital marketer, B2B content strategist, and publisher of the AI Marketing Ethics Digest on Substack. In this episode, Paul joins host John Jantsch to explore the crucial yet overlooked intersection of AI and marketing ethics. From the risks of "shadow AI" and techno-stress to building responsible governance frameworks and his Generative AI Business Adoption Hierarchy, Paul offers a grounded, strategic perspective on how businesses can navigate AI adoption with integrity. Tune in to learn why ethical guardrails aren't just about compliance—they're essential for protecting your brand, your team, and your customers. Today we discussed: 00:09 Introducing Paul Chaney 00:42 Why Paul launched the AI Marketing Ethics Digest 02:58 Transparency, bias, and brand reputation in AI output 05:00 Strategy before technology: avoiding “bad work faster 06:55 What “shadow AI” is and how it can harm organizations 07:55 The need for usage policies and monitoring internal AI use 10:32 The Generative AI Business Adoption Hierarchy explained 13:20 Embedding AI into business culture with governance and clarity 15:27 What is AI techno-stress and how is it impacting workforces? 18:02 Lack of training is a hidden ethical risk for employee well-being 20:08 Why many business owners may give up on AI—and what that means for consultants Rate, Review, & Follow If you liked this episode, please rate and review the show. Let us know what you loved most about the episode. Struggling with strategy? Unlock your free AI-powered prompts now and start building a winning strategy today!

The 10 Minute Teacher Podcast
Teacher Brain Burnout? GPT-4.1 and 8 Other Headlines that Matter

The 10 Minute Teacher Podcast

Play Episode Listen Later May 20, 2025 11:58


Looking for ideas to engage your students in conversation? In this week's quick news roundup, I give you stories about: The impact of overwork on the teacher's brain, An idea for an energy drink experiment for science teachers around the chemical "taurine" NASA and the tectonic plates of Venus YouTube's new "Peak Points" Advertising strategy as an AI article to discuss with students ChatGPT 4 going away? And how I teach students to test different models of AI and share their results How some people are installing local LLM's on their machines New AI guidance for teachers and common patterns I'm noting Google's AIME and the future of medical chat bots DuoLingo goes AI, and A Star-Wars themed personality test gone to the dark side? Once a week, I work to share news articles and stories with you that I'm using. I want you to have quick ideas for turning headlines into a warm-up, debate, or story starter. Show notes and links: https://www.coolcatteacher.com/e902 Sponsor: Rise Vision Do you want to know how I have students share their prompts and test various models of AI? I use my Rise Vision Board! When teaching AI, seeing how each student uniquely interacts with technology is essential. Rise Vision's screen sharing solution turned my aging display into a modern wireless hub without replacement costs. I can now securely moderate which student screens appear—perfect for AI demonstrations and collaborative learning. The Rise Vision system is incredibly user-friendly and costs just a fraction of new interactive displays. I'm saving my school money while enhancing our tech capabilities! Visit Rise Vision to see how you can refresh rather than replace your classroom displays. Link: https://www.coolcatteacher.com/risevision/