Knowledge of means of accomplishing objectives
POPULARITY
Categories
In this episode of Nephilim Death Squad, Raven and Matt Hepner continue The Indiana Zone series with a deep, critical breakdown of the documentary Age of Disclosure.This is not a surface-level reaction. The hosts dissect the UFO disclosure narrative, intelligence community messaging, Vatican involvement, and the spiritual implications behind the modern push to normalize “non-human intelligence.”Topics covered include:Why Age of Disclosure feels engineered, not revealingIntelligence community spokespeople and narrative controlThe emotional and spiritual “heaviness” behind disclosure mediaTransmedium craft, ocean activity, and underground speculationBiological effects, radiation exposure, and military encountersWhy aliens, demons, and deception are being separated on purposeBiblical prophecy, the Days of Noah, and end-times conditioningTimothy Alberino's past warnings vs current disclosure toneThis episode asks the hard question:Is disclosure about truth — or preparation?
Teaser ... What era are we living in? ... David Bowie: "The Internet is an alien life form." ... Bob: "It's like aliens are approaching, and we don't know their intentions" ... Nikita: "COVID was weird, and nothing has been normal since" ... The lab leak theory ... How COVID killed consensus reality ... Bob's case for international biotech monitoring ... Vaccine geopolitics ... Bob's candidate: We let machines guide our attention ... Trump's failed bargaining strategy ... Technological progress is like a cocaine high ... How common is the idea that AI will be a big deal? ... When music is high stakes. Russian street musicians get arrested ... Heading to Overtime ...
Watch every episode ad-free & uncensored on Patreon: https://patreon.com/dannyjones Duncan Trussell is a comedian, writer & host of the @duncantrussellfamilyhour podcast. He also has an animated show on Netflix called “The Midnight Gospel”. Duncan and Danny discuss satan, artificial intelligence, Christianity, Jesus Christ, psychedelics and much more. SPONSORS https://hellofresh.com/danny10fm - Get 10 free meals + a FREE zwilling Knife on your 3rd box. https://www.amentara.com/go/dj - Use code DJ22 for 22% off your first order. https://www.ridge.com/dannyjones - Get up to 47% off on Ridge's biggest sale of the Year. https://whiterabbitenergy.com/?ref=DJP - Use code DJP for 20% off EPISODE LINKS @duncantrussellfamilyhour https://www.duncantrussell.com FOLLOW DANNY JONES https://www.instagram.com/dannyjones https://twitter.com/jonesdanny OUTLINE 00:00 - Preforming at Anton LaVey's satanist wedding 05:47 - Absinthe & Marilyn Manson 12:11 - Dark ages of podcasting 20:53 - Greed 22:54 - Technological singularity 28:41 - Duncan went to Peter Thiel's anti-christ convention 44:44 - AI is a demon that will bring the apocalypse 51:18 - Aleister Crowley & magic spells 57:18 - Christ consciousness 01:06:56 - Link between psychedelics & religion 01:13:59 - The hidden stories of Jesus 01:23:45 - Why they faked the moon landing footage 01:37:37 - How war erases humans' morals 01:44:42 - What a modern apocalypse actually looks like 01:54:43 - Why to delete your social media RIGHT NOW 01:59:09 - Are we in a simulation? 02:01:49 - DMT laser experiment code 02:08:58 - Secret to happiness in 2026 Learn more about your ad choices. Visit podcastchoices.com/adchoices
Following the pace of Jesus and staying interruptible. Challenging the way our culture places efficiency as the highest value. Staying aware of beauty, of goodness, of embodied presence in an age that pushes those into the background.
As early as 1989, intelligence officers in the Defense Intelligence Agency (DIA) recognized China as the next threat, says former DIA officer and physicist Michael Sekora.“We identified what [China] was doing to become a superpower faster than any country in history, and we were on track to containment,” Sekora says.Back in the 1980s, he led a classified Defense Intelligence Agency program called “Project Socrates” that was created under the Reagan administration to determine the cause of U.S. economic and military decline, find a way to reverse it, and outcompete Moscow. Later they turned their sights to Beijing.“It was very obvious what was going on: China was executing a national technology strategy, which basically was playing ... a very adroit game of worldwide offensive, defensive, technology exploitation chess,” Sekora says. “What we had in Socrates could have easily contained China.”The project was defunded by the Bush administration, and the United States went the opposite route, allowing many key technologies to be handed over to Beijing over the course of several decades.In this episode, he breaks down why he believes the United States has lost its edge in technological innovation and how this can be turned around.Views expressed in this video are opinions of the host and the guest, and do not necessarily reflect the views of The Epoch Times.
The Battle of the Atlantic, the longest campaign of World War II, spanned from 1939 to 1945, primarily in the North Atlantic but extending to Arctic convoys and the Mediterranean. This conclusion episode recaps Allied resilience against German U-boats and surface raiders. Britain implemented early rationing and the "Dig for Victory" campaign, increasing food production and reducing import reliance. Key turning points included capturing the Enigma machine from U-110 in 1941, sinking U-boat aces, and U.S. adoption of convoys post-1941. Despite setbacks like Operation Torch diverting escorts in late 1942, commanders like Admiral Max Horton and sub-hunters Donald McIntyre and Frederick "Johnny" Walker turned the tide with aggressive tactics. Technological advances sealed Germany's fate: Allied Hedgehog mortars (far more effective than depth charges), radar, and air cover from B-24 Liberators closed the mid-Atlantic gap by May 1943, forcing U-boat retreats. German innovations like snorkels, Type XXI U-boats, and acoustic torpedoes (e.g., T5 Zaunkönig) arrived too late, yielding high losses (e.g., 39 U-boats for minimal gains in late 1943). In the Arctic, the Scharnhorst's defeat at North Cape secured Soviet supply lines. The episode details D-Day's naval arm (Operation Neptune): Over 6,000 vessels from eight navies, led by Admiral Bertram Ramsay, bombarded Normandy beaches with battleships like HMS Warspite and USS Texas. Despite mines and minor German torpedo boat attacks, it succeeded, ferrying 132,000 troops on June 6, 1944, with Mulberry harbors enabling logistics. The "Hunt for Tirpitz," the Bismarck's sister ship, highlights Allied obsession: Multiple failed raids (e.g., Operation Tungsten) culminated in Operation Catechism (November 1944), where RAF Lancasters dropped 12,000-pound Tallboy "earthquake" bombs, capsizing the "Lonely Queen of the North" and killing up to 1,200 crew. Losses were staggering: Allies lost 3,500 merchant ships (21.6 million tons), 175 warships, and ~7,200 sailors; Germany lost 783 U-boats, 47 warships, and 30,000 men. Submarines sank two-thirds of Allied tonnage, but superior Allied strategy, resources, and innovation prevailed, ensuring victory in this critical supply war.
As technology offers us endless ways to be “present” without actually being there, what does it mean for Christians to imitate the incarnation of Christ? In this episode of Thinking Christian, Dr. James Spencer explores the often-overlooked ethical and theological dimensions of the incarnation—not just that “the Word became flesh,” but how Jesus chose to be with us and for us. Using The Matrix as a modern parable, James contrasts Christ’s self-giving descent with Cypher’s decision to abandon reality and his friends for the comfort of illusion—what James calls “dis-incarnation.” From there, he turns to Philippians 2 to show how Jesus refused to use equality with God for His own advantage, instead embodying a way of life marked by presence, sacrifice, and service. James then examines how dominant ideas like liberalism and transhumanism can subtly train us to embody the world in self-determined, self-serving ways, even when they seem to promise freedom or enhancement. He argues that true Christian embodiment isn’t about maximizing personal options or overcoming our biology through technology, but about conforming our lives to the incarnate Christ—using our gifts, bodies, and opportunities not for our own advantage, but for the good of others. This episode will help you rethink Christmas, technology, and your everyday presence in light of the One who became flesh for us. Subscribe to our YouTube channel
A digital world in relentless movement—from artificial intelligence to ubiquitous computing—has been captured and reinvented as a monoculture by Silicon Valley "big tech" and venture capital firms. Yet very little is discussed in the public sphere about existing alternatives. Based on long-term field research across San Francisco, Tokyo, and Shenzhen, Common Circuits: Hacking Alternative Technological Futures (Stanford UP, 2025) explores a transnational network of hacker spaces that stand as potent, but often invisible, alternatives to the dominant technology industry. In what ways have hackers challenged corporate projects of digital development? How do hacker collectives prefigure more just technological futures through community projects? Luis Felipe R. Murillo responds to these urgent questions with an analysis of the hard challenges of collaborative, autonomous community-making through technical objects conceived by hackers as convivial, shared technologies. Through rich explorations of hacker space histories and biographical sketches of hackers who participate in them, Murillo describes the social and technical conditions that allowed for the creation of community projects such as anonymity and privacy networks to counter mass surveillance; community-made monitoring devices to measure radioactive contamination; and small-scale open hardware fabrication for the purposes of technological autonomy. Murillo shows how hacker collectives point us toward brighter technological futures—a renewal of the "digital commons"—where computing projects are constantly being repurposed for the common good. Mentioned in this episode: "Political Software: Mapping Digital Worlds from Below" Project Website here Luis Felipe R. Murillo is Assistant Professor of Anthropology and Faculty Fellow at the Notre Dame Technology Ethics Center at the University of Notre Dame. His work is dedicated to the study of computing from an anthropological perspective. Liliana Gil is Assistant Professor of Comparative Studies (STS) at The Ohio State University. Learn more about your ad choices. Visit megaphone.fm/adchoices Support our show by becoming a premium member! https://newbooksnetwork.supportingcast.fm/new-books-network
A digital world in relentless movement—from artificial intelligence to ubiquitous computing—has been captured and reinvented as a monoculture by Silicon Valley "big tech" and venture capital firms. Yet very little is discussed in the public sphere about existing alternatives. Based on long-term field research across San Francisco, Tokyo, and Shenzhen, Common Circuits: Hacking Alternative Technological Futures (Stanford UP, 2025) explores a transnational network of hacker spaces that stand as potent, but often invisible, alternatives to the dominant technology industry. In what ways have hackers challenged corporate projects of digital development? How do hacker collectives prefigure more just technological futures through community projects? Luis Felipe R. Murillo responds to these urgent questions with an analysis of the hard challenges of collaborative, autonomous community-making through technical objects conceived by hackers as convivial, shared technologies. Through rich explorations of hacker space histories and biographical sketches of hackers who participate in them, Murillo describes the social and technical conditions that allowed for the creation of community projects such as anonymity and privacy networks to counter mass surveillance; community-made monitoring devices to measure radioactive contamination; and small-scale open hardware fabrication for the purposes of technological autonomy. Murillo shows how hacker collectives point us toward brighter technological futures—a renewal of the "digital commons"—where computing projects are constantly being repurposed for the common good. Mentioned in this episode: "Political Software: Mapping Digital Worlds from Below" Project Website here Luis Felipe R. Murillo is Assistant Professor of Anthropology and Faculty Fellow at the Notre Dame Technology Ethics Center at the University of Notre Dame. His work is dedicated to the study of computing from an anthropological perspective. Liliana Gil is Assistant Professor of Comparative Studies (STS) at The Ohio State University. Learn more about your ad choices. Visit megaphone.fm/adchoices Support our show by becoming a premium member! https://newbooksnetwork.supportingcast.fm/anthropology
A digital world in relentless movement—from artificial intelligence to ubiquitous computing—has been captured and reinvented as a monoculture by Silicon Valley "big tech" and venture capital firms. Yet very little is discussed in the public sphere about existing alternatives. Based on long-term field research across San Francisco, Tokyo, and Shenzhen, Common Circuits: Hacking Alternative Technological Futures (Stanford UP, 2025) explores a transnational network of hacker spaces that stand as potent, but often invisible, alternatives to the dominant technology industry. In what ways have hackers challenged corporate projects of digital development? How do hacker collectives prefigure more just technological futures through community projects? Luis Felipe R. Murillo responds to these urgent questions with an analysis of the hard challenges of collaborative, autonomous community-making through technical objects conceived by hackers as convivial, shared technologies. Through rich explorations of hacker space histories and biographical sketches of hackers who participate in them, Murillo describes the social and technical conditions that allowed for the creation of community projects such as anonymity and privacy networks to counter mass surveillance; community-made monitoring devices to measure radioactive contamination; and small-scale open hardware fabrication for the purposes of technological autonomy. Murillo shows how hacker collectives point us toward brighter technological futures—a renewal of the "digital commons"—where computing projects are constantly being repurposed for the common good. Mentioned in this episode: "Political Software: Mapping Digital Worlds from Below" Project Website here Luis Felipe R. Murillo is Assistant Professor of Anthropology and Faculty Fellow at the Notre Dame Technology Ethics Center at the University of Notre Dame. His work is dedicated to the study of computing from an anthropological perspective. Liliana Gil is Assistant Professor of Comparative Studies (STS) at The Ohio State University. Learn more about your ad choices. Visit megaphone.fm/adchoices Support our show by becoming a premium member! https://newbooksnetwork.supportingcast.fm/science-technology-and-society
A digital world in relentless movement—from artificial intelligence to ubiquitous computing—has been captured and reinvented as a monoculture by Silicon Valley "big tech" and venture capital firms. Yet very little is discussed in the public sphere about existing alternatives. Based on long-term field research across San Francisco, Tokyo, and Shenzhen, Common Circuits: Hacking Alternative Technological Futures (Stanford UP, 2025) explores a transnational network of hacker spaces that stand as potent, but often invisible, alternatives to the dominant technology industry. In what ways have hackers challenged corporate projects of digital development? How do hacker collectives prefigure more just technological futures through community projects? Luis Felipe R. Murillo responds to these urgent questions with an analysis of the hard challenges of collaborative, autonomous community-making through technical objects conceived by hackers as convivial, shared technologies. Through rich explorations of hacker space histories and biographical sketches of hackers who participate in them, Murillo describes the social and technical conditions that allowed for the creation of community projects such as anonymity and privacy networks to counter mass surveillance; community-made monitoring devices to measure radioactive contamination; and small-scale open hardware fabrication for the purposes of technological autonomy. Murillo shows how hacker collectives point us toward brighter technological futures—a renewal of the "digital commons"—where computing projects are constantly being repurposed for the common good. Mentioned in this episode: "Political Software: Mapping Digital Worlds from Below" Project Website here Luis Felipe R. Murillo is Assistant Professor of Anthropology and Faculty Fellow at the Notre Dame Technology Ethics Center at the University of Notre Dame. His work is dedicated to the study of computing from an anthropological perspective. Liliana Gil is Assistant Professor of Comparative Studies (STS) at The Ohio State University. Learn more about your ad choices. Visit megaphone.fm/adchoices Support our show by becoming a premium member! https://newbooksnetwork.supportingcast.fm/technology
What happens when a passion for airplanes meets an unshakeable mission to decarbonize aviation? Kyle Clark the founder and CEO of BETA Technologies explains how electric flight could change the world! Ilham and Klye talk about what it feels like to pilot an electric aircraft, the future of long haul flights, the importance of lightweighting materials, and what makes Beta stand out of its competitors. Did you know that 20% of aviation emissions can be tackled today with existing electric technology?Syensqo is proud to partner with Beta to provide the lightweighting materials for their aircrafts. Kyle Clark is the founder and CEO of BETA Technologies, a company that's redefining sustainable air travel with all-electric aircraft and a scalable approach to certification and commercialization. BETA Technologies is working to make flight greener, safer, and more accessible by helping the world in the transition to electric-powered aviation. BETA builds eVTOL aircraft and charging systems enabling its customers to seamlessly move people and cargo around the world safely, cost-effectively, and with minimal environmental impact. For additional details about the podcast, show notes, and access to resources mentioned during the show, please visit https://www.syensqo.com/en/podcast
https://www.maryharrington.co.uk/p/zombie-nationalism https://www.aaronrenn.com/p/the-elite-that-built-america https://americanaffairsjournal.org/2025/11/the-making-of-a-techno-nationalist-elite/ @FrSteveMacias Abraham Kuyper - Calvinism and the Transformation of Culture - Biography and Impact https://youtu.be/mksoLnBANsE?si=3E6ugazRLRo6Qr9b @WhiteStoneName Paul Alan - Randos #2 & Universal Basic Institutions https://youtu.be/cKohsuEdDAk?si=eds7_LmWgEhpocpU @actoninstitute Abraham Kuyper's Principles for Christian Liberalism with Matthew Tuininga https://youtu.be/gNgdItCGMUc?si=XRMdrbis4qy9CSLV @TheAaronRennShow What Killed America's Can-Do Culture? | Tanner Greer https://youtu.be/V1cwOW9l2L4?si=zf0unwq1LTb-8qYq https://scholars-stage.org/book-notes-the-technological-republic-2025/ https://www.livingstonescrc.com/give Register for the Estuary/Cleanup Weekend https://lscrc.elvanto.net/form/94f5e542-facc-4764-9883-442f982df447 Paul Vander Klay clips channel https://www.youtube.com/channel/UCX0jIcadtoxELSwehCh5QTg https://www.meetup.com/sacramento-estuary/ My Substack https://paulvanderklay.substack.com/ Bridges of meaning https://discord.gg/mQGdwNca Estuary Hub Link https://www.estuaryhub.com/ There is a video version of this podcast on YouTube at http://www.youtube.com/paulvanderklay To listen to this on ITunes https://itunes.apple.com/us/podcast/paul-vanderklays-podcast/id1394314333 If you need the RSS feed for your podcast player https://paulvanderklay.podbean.com/feed/ All Amazon links here are part of the Amazon Affiliate Program. Amazon pays me a small commission at no additional cost to you if you buy through one of the product links here. This is is one (free to you) way to support my videos. https://paypal.me/paulvanderklay Blockchain backup on Lbry https://odysee.com/@paulvanderklay https://www.patreon.com/paulvanderklay Paul's Church Content at Living Stones Channel https://www.youtube.com/channel/UCh7bdktIALZ9Nq41oVCvW-A To support Paul's work by supporting his church give here. https://tithe.ly/give?c=2160640 https://www.livingstonescrc.com/give
PREVIEW — Chris Riegel — "The Great Heist": China's Intellectual Property Theft Strategy. Riegel argues that Chinahas systematically employed intellectual property theft as a strategic mechanism to advance from technological desperation toward status as a near-peer American rival, systematically acquiring technologies while circumventing legitimate licensing and development pathways. Riegel documents that American corporations frequently ignored or tolerated Chinese IP theft in exchange for access to the lucrative Chinese domestic market, including the Cisco-Huaweicase illustrating corporate preference for market access over intellectual property protection. Riegel acknowledges that China now conducts original research and technological innovation, yet systematically rejects international intellectual property law regimes and refuses to recognize or enforce IP rights, maintaining institutional resistance to the legal frameworks protecting legitimate innovation and development within Western economies. 1922
This episode explores the high-stakes balancing act logistics companies face as they navigate internal optimization against a backdrop of external chaos. We begin by analyzing the sharp drop in benchmark diesel prices as domestic refineries ramp up production, though experts warn of a potential oil super glut by 2026. In the cold chain sector, Lineage is combatting excess capacity by rolling out its LinOS system to generate $110 million in efficiency gains through automation. Meanwhile, the automotive logistics space sees strategic consolidation as RPM acquires PARS to create a comprehensive vehicle lifecycle platform that includes driveaway and titling services. Technological advancements continue as Daimler Truck and Torc Robotics integrate short-range LiDAR to bring autonomous freightliners to the North American market. However, carriers must remain vigilant on compliance, as the FMCSA has revoked 62% more ELDs this year, threatening immediate shutdowns for fleets using banned devices. Corporate governance takes center stage with the affirmed 20-year sentence for Slync.io's founder, serving as a stark reminder of the importance of fiduciary trust in freight tech. Finally, we examine the supply chain risks emerging from threatened tariffs on Mexico tied to a dispute over water deliveries that is already impacting Texas agriculture. Learn more about your ad choices. Visit megaphone.fm/adchoices
I'm old enough to remember when ABC, NBC and CBS dominated our television viewing and our consumption of electronic news, and that the viewer had to be in front of the television set at an appointed time to receive the information. Today, the gatekeeper, who once decided when and what type of news content … Read More Read More
This episode explores the high-stakes balancing act logistics companies face as they navigate internal optimization against a backdrop of external chaos. We begin by analyzing the sharp drop in benchmark diesel prices as domestic refineries ramp up production, though experts warn of a potential oil super glut by 2026. In the cold chain sector, Lineage is combatting excess capacity by rolling out its LinOS system to generate $110 million in efficiency gains through automation. Meanwhile, the automotive logistics space sees strategic consolidation as RPM acquires PARS to create a comprehensive vehicle lifecycle platform that includes driveaway and titling services. Technological advancements continue as Daimler Truck and Torc Robotics integrate short-range LiDAR to bring autonomous freightliners to the North American market. However, carriers must remain vigilant on compliance, as the FMCSA has revoked 62% more ELDs this year, threatening immediate shutdowns for fleets using banned devices. Corporate governance takes center stage with the affirmed 20-year sentence for Slync.io's founder, serving as a stark reminder of the importance of fiduciary trust in freight tech. Finally, we examine the supply chain risks emerging from threatened tariffs on Mexico tied to a dispute over water deliveries that is already impacting Texas agriculture. Learn more about your ad choices. Visit megaphone.fm/adchoices
FAITHBUCKS.COM
REVIEW: Breakthrough in Ground-Based Exoplanet Imaging: Colleague Bob Zimmerman highlights a technological breakthrough where astronomers used the ground-based Subaru telescope in Hawaii to image a "super Jupiter" exoplanet 271 light-years away, successfully capturing the planet orbiting by blocking the host star's light, a feat demonstrating amazing advances in astronomical software and observational technology.
Support Us: Donation Page – LibriVox Free AudiobooksAnthem is a dystopic science fiction story taking place at some unspecified future date. Mankind has entered another dark age as a result of what Rand saw as the weaknesses of socialistic thinking and economics. Technological advancement is now carefully planned (when it is allowed to occur, if at all) and the concept of individuality has been eliminated (for example, the word "I" has disappeared from the language). As is common in her work, Rand draws a clear distinction between the "socialist/communal" values of equality and brotherhood and the "productive/capitalist" values of achievement and individuality. The story also parallels Stalinist Russia, which was currently going on at the time as the story was published. (Summary from Wikipedia)Support Us: Donation Page – LibriVox Free Audiobooks
Matthew 12:1 At that time Jesus went on the sabbath day through the corn; and his disciples were an hungred, and began to pluck the ears of corn, and to eat. 12:2 But when the Pharisees saw it, they said unto him, Behold, thy disciples do that which is not lawful to do upon the sabbath day. 12:3 But he said unto them, Have ye not read what David did, when he was an hungred, and they that were with him; 12:4 How he entered into the house of God, and did eat the shewbread, which was not lawful for him to eat, neither for them which were with him, but only for the priests? 12:5 Or have ye not read in the law, how that on the sabbath days the priests in the temple profane the sabbath, and are blameless? 12:6 But I say unto you, That in this place is one greater than the temple. 12:7 But if ye had known what this meaneth, I will have mercy, and not sacrifice, ye would not have condemned the guiltless. 12:8 For the Son of man is Lord even of the sabbath day. 12:9 And when he was departed thence, he went into their synagogue: 12:10 And, behold, there was a man which had his hand withered. And they asked him, saying, Is it lawful to heal on the sabbath days? that they might accuse him. 12:11 And he said unto them, What man shall there be among you, that shall have one sheep, and if it fall into a pit on the sabbath day, will he not lay hold on it, and lift it out? 12:12 How much then is a man better than a sheep? Wherefore it is lawful to do well on the sabbath days. 12:13 Then saith he to the man, Stretch forth thine hand. And he stretched it forth; and it was restored whole, like as the other. 12:14 Then the Pharisees went out, and held a council against him, how they might destroy him.
Technological development has always been a double-edged sword for humanity: the printing press increased the spread of misinformation, cars disrupted the fabric of our cities, and social media has made us increasingly polarized and lonely. But it has not been since the invention of the nuclear bomb that technology has presented such a severe existential risk to humanity – until now, with the possibility of Artificial Super Intelligence (ASI) on the horizon. Were ASI to come to fruition, it would be so powerful that it would outcompete human beings in everything – from scientific discovery to strategic warfare. What might happen to our species if we reach this point of singularity, and how can we steer away from the worst outcomes? In this episode, Nate is joined by Nate Soares, an AI safety researcher and co-author of the book If Anyone Builds It, Everyone Dies: Why Superhuman AI Would Kill Us All. Together, they discuss many aspects of AI and ASI, including the dangerous unpredictability of continued ASI development, the "alignment problem," and the newest safety studies uncovering increasingly deceptive AI behavior. Soares also explores the need for global cooperation and oversight in AI development and the importance of public awareness and political action in addressing these existential risks. How does ASI present an entirely different level of risk than the conventional artificial intelligence models that the public has already become accustomed to? Why do the leaders of the AI industry persist in their pursuits, despite acknowledging the extinction-level risks presented by continued ASI development? And will we be able to join together to create global guardrails against this shared threat, taking one small step toward a better future for humanity? (Conversation recorded on November 11th, 2025) About Nate Soares: Nate Soares is the President of the Machine Intelligence Research Institute (MIRI), and plays a central role in setting MIRI's vision and strategy. Soares has been working in the field for over a decade, and is the author of a large body of technical and semi-technical writing on AI alignment, including foundational work on value learning, decision theory, and power-seeking incentives in smarter-than-human AIs. Prior to MIRI, Soares worked as an engineer at Google and Microsoft, as a research associate at the National Institute of Standards and Technology, and as a contractor for the US Department of Defense. Show Notes and More Watch this video episode on YouTube Want to learn the broad overview of The Great Simplification in 30 minutes? Watch our Animated Movie. --- Support The Institute for the Study of Energy and Our Future Join our Substack newsletter Join our Hylo channel and connect with other listeners
The Evolution of Drone Warfare in Ukraine — John Batchelor, Bill Roggio, John Hardie — Roggio examines drone warfare's significant but sometimes exaggerated role in the Ukraine conflict, tracing technological evolution from ISRreconnaissance drones like the TB2 to FPV (First Person View) drones, bomber drones, and emergent interceptor drone systems. Hardie identifies a consistent pattern: Ukraine consistently leads in innovative military technology development and deployment, while Russia demonstrates superior capability in rapidly scaling technologies to mass production and tactical integration, occasionally achieving operational effectiveness superior to the original Ukrainian innovations. Batchelor emphasizes this dynamic reflects fundamentally different military cultures and organizational capabilities.
In this episode of All in the Delivery, they bring on retired driver and Circle of Honor member of 25 years safe driving Vic Wirhowski to discuss the evolution of UPS, focusing on the changes in company culture, service quality, and appearance standards over the years. They share personal anecdotes and reflect on how the industry has shifted towards profit-driven priorities, often at the expense of safety and service. The conversation highlights the gradual decline in standards and professionalism, as well as the impact of technology and cultural shifts on the workforce. In this conversation, the speakers reflect on their experiences as UPS drivers, discussing the evolution of the job, the impact of technology, and the changes in company culture over the years. They share personal stories about navigating challenges, including pension cuts and the restoration of benefits, while emphasizing the importance of recognition and morale in the workplace. The discussion highlights the contrast between past and present experiences, illustrating how the driving profession has transformed and the lasting legacy of service within the company. Help support the show Join our Discord Takeaways The importance of acknowledging supporters and community. Personal stories can illustrate broader industry issues. Career changes can bring a sense of relief and fulfillment. Company changes often lead to a decline in service quality. Profit-driven decisions can compromise safety and service. The deterioration of standards is often gradual and unnoticed. Appearance standards were once strictly enforced in the industry. Technological advancements have changed operational practices. Cultural shifts have influenced company policies on appearance. The balance between evolution and maintaining standards is crucial. The job has become easier due to technology like GPS. Pension cuts were a significant concern for retirees. Restoration of pensions brought relief and gratitude. Company morale has declined with fewer recognition events. Drivers used to feel appreciated with regular rewards. The culture of the workplace has shifted over the years. Recognition for safe driving has diminished. Corporate messages often focus on safety outside of work. There are always employees who resist hard work. Legacy and service are core values for long-term employees. Chapters 00:00 Introduction and Acknowledgments 01:49 Personal Stories and Experiences 06:18 New Beginnings and Routes 08:52 Company Changes and Service Quality 13:03 The Slow Deterioration of Standards 17:35 Impact of COVID-19 on Operations 20:16 Leadership Changes and Their Effects 21:37 The Evolution of Driver Appearance Standards 24:32 Technological Advancements in Delivery 30:45 Changes in Union Regulations and Pension 37:44 Reflections on Job Expectations and Experience 44:08 Pension Cuts and Morale Issues 45:16 The Importance of Morale in the Workplace 46:52 Challenges with Team Dynamics 48:28 Recognition and Appreciation in the Workplace 50:27 The Shift in Company Culture 52:10 Technical Difficulties and Communication 53:57 Reflections on Union and Safety 56:27 Founders Day and Company Traditions 59:03 Positive Perspectives on Company Experience THE OPINIONS EXPRESSED OR VIEWS EXPRESSED ON THIS PODCAST ARE THOSE OF THE HOSTS AND GUESTS AND DO NOT NECESSARILY REFLECT ANY DELIVERY COMPANY
Episode 143I spoke with Iason Gabriel about:* Value alignment* Technology and worldmaking* How AI systems affect individuals and the social worldIason is a philosopher and Senior Staff Research Scientist at Google DeepMind. His work focuses on the ethics of artificial intelligence, including questions about AI value alignment, distributive justice, language ethics and human rights.You can find him on his website and Twitter/X.Find me on Twitter (or LinkedIn if you want…) for updates, and reach me at editor@thegradient.pub for feedback, ideas, guest suggestions.Outline* (00:00) Intro* (01:18) Iason's intellectual development* (04:28) Aligning language models with human values, democratic civility and agonism* (08:20) Overlapping consensus, differing norms, procedures for identifying norms* (13:27) Rawls' theory of justice, the justificatory and stability problems* (19:18) Aligning LLMs and cooperation, speech acts, justification and discourse norms, literacy* (23:45) Actor Network Theory and alignment* (27:25) Value alignment and Iason's starting points* (33:10) The Ethics of Advanced AI Assistants, AI's impacts on social processes and users, personalization* (37:50) AGI systems and social power* (39:00) Displays of care and compassion, Machine Love (Joel Lehman)* (41:30) Virtue ethics, morality and language, virtue in AI systems vs. MacIntyre's conception in After Virtue* (45:00) The Challenge of Value Alignment* (45:25) Technologists as worldmakers* (51:30) Technological determinism, collective action problems* (55:25) Iason's goals with his work* (58:32) OutroLinksPapers:* AI, Values, and Alignment (2020)* Aligning LMs with Human Values (2023)* Toward a Theory of Justice for AI (2023)* The Ethics of Advanced AI Assistants (2024)* A matter of principle? AI alignment as the fair treatment of claims (2025) Get full access to The Gradient at thegradientpub.substack.com/subscribe
In this episode of the WB Download, host Jeff Wieland takes listeners on a journey through the evolution of the home building industry from the days when his father, Jack Wieland, owned Wieland Builders, to today, where Jeff leads the company alongside his brother, Mike Wieland.Jeff reflects on decades of transformation, highlighting how technology, tools, materials, and even communication have reshaped the way custom homes are built.Technological & Industry AdvancesJeff looks back at the early days of landlines, handwritten messages, and typewriters, contrasting them with today's cell phones, digital pagers, and computers that streamline communication and planning.He shares how job-site equipment has progressed from front-engine loaders to modern rear-engine loaders and track hoes, along with the introduction of concrete pumps and gravel slinger trucks that make construction safer and more efficient.Changes in Building Practices & MaterialsJeff breaks down the major shifts in materials and methods, including:Moving from single-pane windows to highly efficient insulated glassThe introduction of house wrap and air infiltration barriers for better moisture and energy controlImproved plumbing fixtures, pressure regulators, updated lighting, and the rise of LED technologyEnhanced insulation options transitioning from open-faced insulation to craft-backed insulation and today's spray foam solutions He also talks about how the industry moved from in-house labor to a greater reliance on specialized subcontractors.Safety & RegulationsJeff discusses OSHA's growing impact on job-site safety and how new standards have influenced equipment choices, from safer ladders to improved extension cords.Code updates, such as requiring house wrap and continuous gutters have pushed builders toward higher quality and energy-efficient practices.Hiring & Marketing in a Modern WorldFrom newspaper job ads to online platforms like Indeed, Jeff explains how hiring has changed dramatically.He also highlights the importance of marketing and social media for showcasing their craftsmanship and connecting with customers, something that didn't exist in Jack's era.Gratitude, Reflections & What's AheadJeff shares his appreciation for the many positive changes in materials and building practices, even those driven by manufacturers and expresses gratitude for the continued support of podcast listeners.He wraps up the episode by wishing everyone a Happy Thanksgiving and reminding them to visit their website for more insights into their projects.Email Jeff your comments, questions, and topic requests, or be a guest on The WB Download.Email: WBDOWNLOAD@wielandbuilders.comSee Wieland Builders custom home gallery www.wielandbuilders.comReceive inspiration monthly in our monthly newsletter See podcast behind the scenes photosFollow us on Facebook, Instagram, YouTube, Houzz or Pinterest
Send us a textDr Alex Curmi is a psychiatrist and psychotherapist who also hosts The Thinking Mind podcast, and is a gifted communicator on mental health and self-development. Alex's clinical work and training has given him acute insights into troubling aspects of modern life, and how we might prepare for an uncertain future. The question which formed the spine of our conversation was ‘ In a world where technology has been quite disruptive psychologically for a lot of people, how do we prepare for an increasingly technological future?We examine how modern technology reshapes attention, confidence, morality and meaning, and Alex offers practical tips for staying human as machines grow more capable. Among the topics explored you will find:• tech-driven overstimulation dulling joy and focus• confidence built through voluntary discomfort• psychiatry and psychotherapy as complementary lenses• intolerance of uncertainty and stoic control• integrity, congruence and moral habits that scale• social skills as a proactive practice• AI as tool versus thinking crutch• career durability through uncommon skill stacks• financial resilience over consumerist drift• community as the container for lasting changeIf you do enjoy the show, please follow or rate it. It really helps others to find it. For future episodes and news on the show, please sign up to the substack - https://thisexaminedlife.substack.com/?utm_campaign=profile_chipsSupport the show
The following is AI-generated approximation of the transcript from the Equipping Hour session. If you have questions you would like to be addressed in followup sessions, please direct those to Jacob. Opening & Introduction Smedly Yates: All right, this morning’s equipping hour will be about artificial intelligence—hopefully an attempt to introduce this topic, help us think through it carefully, well, biblically. Let me just open our time in prayer. [Prayer] Heavenly Father, thank you so much for your kindness to us. Thank you for giving us all that we need for life and godliness, for not leaving your people adrift. Thank you for putting us into this world exactly in the era that you have. We pray to be effective, fruitful, in all those things which matter for eternity in this world, in this time, in this age. God, we pray for wisdom, that you would guide our discussion here. We pray that this would be of benefit and a help to Grace Bible Church. We ask it in Jesus’ name. Amen. Here’s the layout for this morning and for a future equipping hour. We’ll be talking for about 35 minutes, back and forth—Jake and I—and then at 9:35, the plan is to go to Q&A. So, this is an opportunity for you to ask questions. At that point, I’ll surrender my microphone and you guys can rove and find people. For the next 33 minutes or so, you can be thinking about the questions you’d like to ask. Jake’s going to do most of the talking in our time here. I’m going to set him up with some questions, but just by way of intro, I want to get some things out of the way as we’re talking about artificial intelligence. You might be terrified, you might be hopeful. I want to get the scary stuff out of the way first and tell you what we’re not going to talk about this morning. Is that fair? Artificial intelligence is here. Some of you are required to use it in the workplace. Some of you are prohibited from using it in your workspaces. There’s nothing you and I can do to keep it from being here. Some of the dangers, some of the things you might be wondering about, some of the things that make the news headlines—over the last two weeks, scanning the headlines, there was a new AI headline every day. One of the terrible things that we won’t talk about today is the fact that nobody knows what’s true anymore, right? How can we discern? But the reality is the god of this world has been Satan for the entirety of human history and he’s a deceiver from the beginning. There’s nothing new about lies. They might be easier and more convincing with certain technological advances. The lies might be more ubiquitous, but the same humanity and the same satanology are at play. We may be concerned about societal fracture and distrust. Some people, if they distrust new tech, will withdraw from society. Others will fully embrace it. And so you get a fracture in society—those with, and those without tech. Some people will just say, “If the digital world works, we’re going to use it.” That’s not the Christian perspective. We’re not simply pragmatists. We do care about what’s true and what’s right. Some are worried about AI chatbot companions that will mark the extinction of relationships, marriage, society. I probably fall into the category of those who assume that AI will mean the end of music or the death of music and other art forms. That’s just me, a confession. People run to end-of-the-world scenarios—the robots decide they don’t need us anymore or the collective conscience of AI decides that humanity is a pollutant on Mother Earth, and the only way to keep the earth going is to rid itself of humanity. The survival of the planet is dependent on our own extinction. So AI will bring about a mass human genocide and the end of homo sapiens on earth. We know that’s not true, right? We know how the world ends, and it doesn’t end by an AI apocalypse. So don’t worry about that. Some people worry that AI will be a significant civilization destabilizer. That might be true. But we know that God is sovereign, and we know where society and civilization end up: at the feet of Jesus worshipping him when he rules on the earth for a thousand years leading into the eternal state. So don’t worry about that either. Some believe that AI is the antichrist. Now we know that’s not true. What is the number of the beast? 666. And this year it got rounded up to 67. So we know AI is not the antichrist. 67 is the antichrist. And if you want to know why the numbers six and seven got together in the year 2025 and formed the new word of the year, ask your middle schooler. Is that all the scary stuff? Not even close. I have a family member who has worked in military intelligence working on artificial intelligence stuff for a long time. He said it’s way scarier than you could possibly imagine. Do you want to say any more other scary scenarios we shouldn’t be thinking about? Jacob Hantla: No, we’ll probably cover some of those. Smedly Yates: Okay, great. What we want to focus on today is artificial intelligence as a tool. Just as an axe can be a tool for good or evil, AI is a tool that either has opportunities for betterment or opportunities for danger. So we want to think about that well. What you have on stage here are two of the shepherds at Grace Bible Church. You’ve got Jake Hantla, who is the guy I want exploring artificial intelligence and telling us how to use it well—he has and he does. And then you have me; I intend not to use artificial intelligence for now. We’re on opposite ends of a spectrum, but we share the same theology, same principles, same concerns, and I think the same inquisitive curiosity about technological advances. I drive a car; I’m not Amish in a horse and buggy. I like tech. But on this one, I’m just going to wait and see. I’m going to let Jake explore. From these two different poles, I hope we can be helpful this morning to help us all together think through artificial intelligence. What is AI? Smedly Yates: Let’s start with this, Jake. What is AI basically? Jacob Hantla: At the heart of it, most forms of AI are a tool to predict the next token. That might not mean much to you, but it’s basically a really fancy statistical prediction machine that accomplishes a lot of really powerful outcomes. It doesn’t have a mind, emotions, or consciousness, but it can really effectively mimic those things because it’s been trained on basically all that humanity has produced that’s available to it on the web and in other sources. I’ll try not to be super technical, but I want to pop up a picture. Can you go to slide one? When we think of AI, large language models are probably the one that most of you will think of: ChatGPT, Gemini, Grock, Claude, things like that. Effectively, what it does when we’re thinking of language—it can do other things, like images and driving cars and other things, but let’s think of words—it takes basically all that humanity has written and learns to predict the next token, or we could just think of the next word. So, all of you know, if I said, “Paris is a city in…” most of you would say France. Paris is a city in France. How do you know that? Everyone here has learned that fact. Large language models have gone through a process of training where they learn facts, concepts, and grammar, so that they can effectively speak like a human in words, sentences, and paragraphs that make sense. So how did it get to that? On the right, there’s just a probability that “France” is the most probable next word. How did it get there? Next slide. I’ll go fast. Basically, it’s a whole bunch of tunable weights—think of little knobs or statistical probabilities that interlink parameters. These things get randomized—there are trillions of them in the modern large language models. They’re just completely random, and then it starts feeding in text. Let’s say it was “It was the best of times, it was the…” and it might say “gopher” as the next word when you just randomly start, and that’s obviously wrong. The right word would be “worst.” So, over and over and over again, for something that would take one computer about a hundred million years to do what they do in the pre-training, they have lots of computers doing this over and over until it can adequately say, “Nope, it wasn’t gopher. It should be worst. Let’s take another crack at it.” It just manipulates these knobs until it can act like a human. If you fed it a mystery novel and at the end it would say, “The killer was…” it has to be able to understand everything before to adequately guess who the killer was, or “What is the capital of France?” It compresses tons and tons of knowledge from all of the written text. Then you start putting images in and it compresses knowledge from images and experience from life into a whole bunch of knobs—basically, numbers assigned so it can have an output that is reasonable. Next slide. You take people—pre-training is the process where you’re basically feeding text into it and it’s somehow learning. We don’t even know—humans are not choosing which knobs mean what. It’s a black box. We can sort of start to figure out which knobs might mean things like masculinity or number or verbs, but at the end, you just have a big bunch of numbers. Then humans come in and train it—reinforcement learning with human feedback. They say, “This is the kind of answers we want this tool to give.” At the outcome, people are saying, “We ask it a question, it outputs an answer, we say that’s a good one, that’s a bad one.” But in this, you can see there’s lots of opportunity for falsehood or biases—unstated or purposeful—to sneak in. If you feed in bad data into the training set, and if it’s trained on all of the internet—all that humans have made—you’re going to have a whole lot of truth in there, but also a whole lot of falsehood. It’s not learning to discern between those things; it’s learning all those things. In reinforcement learning with human feedback, we’re basically fine-tuning it, saying, “This is the kind of answer we want you to give,” and that’s going to depend on who teaches it. Then the final step is people judging the answers: “This is the kind of answer we want, this is the kind we don’t want.” Lots of opportunity for biases to sneak in. That was a long answer to “What is AI?” It’s a prediction machine with a whole lot of math going on. What Sets AI Apart from Other Technology? Smedly Yates: Jake, what sets AI apart from previous technological advances, especially as it relates to intention? Jacob Hantla: Tech could be as simple as writing, the wheel, the airplane, telephones, the internet—all those things. All of those, in some sense, enhanced human productivity, strength, our ability to communicate. We could pick up a phone and communicate over distance, use radio waves to communicate to more people, but it was fundamentally something that humans did—magnified. A tractor takes the human art, the human attempt to cultivate a field, and increases efficiency. AI can actually do that. A human in control of an AI can really augment the productivity and effectiveness of a human. You could read a book yourself to gain knowledge or have AI read a book, summarize it, and you get the knowledge. But AI can, for the first time, generate things that look human. It’s similar in some ways, but it’s very different in that it’s generative. AI and Truth Smedly Yates: Tell me about the relationship between AI and truth. You touched on it a little bit before. Jacob Hantla: AI contains a lot of truth. It’s been trained on even ultimate truth. AI has read the Bible more times than any of us ever could. To a large degree, it understands—as AI can understand—a lot of true things and can hold those truths simultaneously in ways that we can’t. But mixed in is a lot of untruth, and there’s no… AI can’t have the Holy Spirit. AI isn’t motivated the same way we are to know what’s true, to know what’s not. So, AI contains a lot of truth and can help you get to truth. You can give it a bunch of true documents and say, “Can you help me? Can you summarize the truth that’s in here? Or actually just summarize what’s in here?” If what’s in there was true, the output will be true; if what’s in there was false, it will output falsehood. It doesn’t have the ability or the desire to determine what is true and what’s not. AI, Emotion, Values, and Worldview Smedly Yates: So, ability and desire are interesting words. Let’s talk about emotion in AI, values in AI, worldview, and regulation of data. For us, true/false claims matter—or they don’t—depending on our worldview and values. Is there a mystery inside this black box of values, of emotion? How do we think about that? Jacob Hantla: First, AI doesn’t inherently have emotion or values, but it can mimic it based on the data it’s been trained on. You can ask the same AI a question and, unless you guide it, it will give you likely a hundred different answers if you ask the same question a hundred times. Unless it’s been steered in one direction, some answers will be good, some will be bad—everything in between. It’s generating a statistical probability. It doesn’t inherently have any of those things but can mimic them. It can be trained to have the values of the trainers. You can have system prompts where the system is prompted to respond in a way that mimics values, mimics emotions. The danger is if you just accept what it says as truth, which a lot of people will do. You say, “I want to know a piece of data,” and you ask the AI and the answer comes out, and you accept it. But you have to understand the AI is just generating a response based on probabilities. If you haven’t guided it to have a set of values, you don’t know what’s going to come out—and somebody may hide some values in it. Gemini actually did this. I think it was Gemini 2, but if you asked for a picture of the Founding Fathers, it would—because it was taught in the system prompt to prioritize diversity—give you images of a diverse group of females or different races, other than the races of the actual Founding Fathers, because it was taught to prioritize that. It had a hidden value in it. You can guide it to have the values you want with a prompt. It’s not guaranteed, but this is the kind of thing I would encourage you to do if you’re using these tools: put your own system prompt on it, tell it what worldview you want it to come from, what your aim is, and you’ll get a more helpful answer than not. Is AI Avoidable? Smedly Yates: Is AI something we can avoid, ignore, be blissfully ignorant about, put our heads in the sand? Jacob Hantla: You could, but I think it’s wise that we all think about it. I’m not encouraging people to adopt it in the same way that I have or Smed has. But the reality is, the world around us has changed. It’s irreversibly different because of the introduction of this technology. That’s what happens with any technology—you can’t go back. Technological advances are inevitable, stacked from scientific discovery and advances. If OpenAI wasn’t doing what it’s doing, somebody else would. You can’t go back. You can’t ignore it because the world is going to be different. You’re going to be influenced by both the presence of it and the output of it. When you get called on the phone now with a very believable voice, it might not be the person it sounds like—AI can mimic what it’s been trained on. There’s thousands of hours of Smed’s voice; it won’t be long before Smed could call you and it’s not Smed. Or Scott Demerest could send you an email asking for a credit card and it’s not Scott. News reports are generated by AI; some of them are true, effective, good summaries, and some could be intentionally spreading disinformation or straight-up falsehood. If you’re not aware of the presence of these things, you could be taken advantage of. Some work environments now require you to do more than you could have otherwise, and not being willing to look at the tools in some jobs will make you unable to compete. Commercially Available AI Products: Benefits and Dangers Smedly Yates: Let’s talk about the commercially available AI products that people can access as a tool. What are the opportunities, the benefits, and what are some of the dangers? Jacob Hantla: There are so many we couldn’t begin to go through all of them, but the ones most of you will interact with are large language models—people just say “ChatGPT” like Kleenex for tissues. It was the first one that came out and is probably the most ubiquitous, one of the easiest to use, and most powerful free ones. There’s ChatGPT by OpenAI, Gemini by Google, Claude by Anthropic, Grock by X.AI (Elon Musk’s), DeepSeek from China (good to know that’s made/controlled by China), Meta’s Llama, etc. Do the company names matter? Yes. It’s good to know who made it and what their goals are, because worldviews are to some degree baked into the model. If you’re ignorant of that, you’ll be more likely to be deceived or not use the tool to the maximum. But with all of these, these are large language models. I drive around now with AI driving my car—ultimately, it’s a similar basis, but that’s not our focus here. Large language models open up the availability of knowledge to us. They’re superpowered Google searches. You can upload a bunch of journal articles, ask it to train you to mastery on a topic. For example, I was trying to understand diastolic heart failure and aortic stenosis—uploaded articles, had a built-in tutor. The tutor asked me questions, evaluated my understanding, used the Socratic method to train me to mastery. This could do in 45 minutes what would have taken me much longer on my own. Every tool can do that. The bad side: you could have it summarize articles for you, and now feel like you have mastery you didn’t actually gain. You could generate an essay or pass a test using it, bypassing the entire process of learning and thinking. Students: if you have a tool that mimics human knowledge and creativity, and you have an assignment to write an essay, and you turn in what the tool generated as your own, you’re being dishonest and you bypass the learning process. The essay wasn’t the point—the process was. Passing a test is about assessing if you know things. If the AI does it for you, you bypass learning. I liken it to going to the gym. The point isn’t moving the weights, it’s building muscle. With education, the learning process is like exercise. It’s easy to have AI do the heavy lifting and think you did it, but you didn’t get stronger. So, be aware of what you’re losing and what you’re gaining. The tool itself isn’t morally good or bad; it’s how the human uses it. The more powerful the technology, the greater good or evil can be accomplished. The printing press could distribute Bibles, but also propaganda. Using AI with Worldview and Preferences Jacob Hantla: When I interact with AI on the Bible, I put a prompt: “When I ask about the Bible or theology, you will answer from a conservative, evangelical, Bible-believing perspective that uses a literal, grammatical-historical hermeneutic and a premillennial eschatology. Assume the 66-book Protestant canon is inspired, inerrant, infallible, completely trustworthy, without error in the original manuscripts, sufficient, and fully authoritative in all it affirms. No sources outside of the 66 books of this canon should be regarded as having these properties. Truth is objective, not relative; therefore, any claim that contradicts the Bible so understood is wrong.” I’m teaching it to adopt this worldview. If you don’t set your preferences, you might get any answer. The tool can learn your preference over time, but it’s better to set it explicitly. Audience Q&A Presuppositions and Biases in AI Audience (Nick O’Neal): What about the values and agenda behind those who input the data? What discernment do the programmers have to put that information in? Jacob Hantla: That goes to baked-in presuppositions or assumptions in the model. Pre-training is basically non-discerning: it’s huge chunks of everything ever written—good, bad, ugly, in between. It’s trained not on a set of values. Nobody programs values in directly; the people making it don’t even know what's being baked in. The fine-tuning comes when trainers judge outputs and reinforce certain responses. System prompts—unseen by users—further guide outputs, reflecting company worldviews. Companies like OpenAI are trying to have an open model so each person can let it adopt their own worldview, but there are still baked-in biases. For example, recent headlines showed some models valuing certain people groups differently, which reflects issues in training data or the trainers' worldview. You’re right to always ask about the underlying assumptions, which is why it would be foolish to just accept whatever comes out as truth. In areas like engineering, worldview matters less, but in many subjects, the biases matter. Is There an AI Bubble? Audience (Matthew Puit): When AI came out, the costs rose artificially by companies. Is the AI bubble going to pop? Jacob Hantla: I don’t know. I think AI will be one of the most transformational technologies. It’ll change things in ways we anticipate and in ways we don’t. Some people will make a lot of money, some will flop. If I knew for sure, I could make a lot of money in the stock market. AI-Generated Worship Music Audience (Rebecca): I see AI-generated worship music based on Psalms, but it’s generated by AI. Is anything lost in AI-generated worship music? Jacob Hantla: AI doesn’t have a soul or the Holy Spirit. It can generate worship music with good doctrine, but that doctrine didn’t come from a place of worship. AI can pray a prayer, but the words aren’t the result of a worshipful heart. You can worship God with those words, but you’re not following a human author who was worshipping God. For example, my kids used Suno (an AI music tool) to set a Bible verse to music for memorization—very helpful. Some might be uncomfortable with music unless it was created by a human; that’s a preference. Creativity is changing, and it will get hard to tell if music or video was made by a human or by AI. That distinction is getting harder to make every day. Setting Preferences in AI Tools Audience (Lee): You mentioned putting your preferences in. How do I do that, especially with free tools? Jacob Hantla: Paid AIs get more processing power, context window, and can use your preferences more consistently. Free versions have some ability—you can usually add preferences in the menu. But even if not, you can paste your preferences at the beginning of your question each time: define who you are, what you want, what worldview to answer from. For example: “I’m a Bible-believing Christian,” or “I’m a nurse anesthesiologist.” That helps the AI give a better answer. Parental Guidance and Children Using AI Smedly Yates: What should parents be aware of in helping their kids navigate AI? Jacob Hantla: Be aware of dangers and opportunities. Kids will likely use these tools, so set limits and help them navigate well. These tools can act like humans—kids without friends might use them as companions, and companies are adding companion avatars, some with sinful tendencies. That can be a danger. For school, a good use is as a tutor: after a quiz, have your child upload the results and ask, “Help me understand where I’m weak on this topic.” But also, be aware of the temptation to use AI to cheat or shortcut the process of learning, discovery, and thinking. Which AI Model? Will AI Become Self-Aware? Audience (Steve): Is there a model you recommend? And does the Bible preclude the possibility of AI becoming self-aware? Jacob Hantla: There’s benefits and drawbacks to all. For getting started, ChatGPT or Perplexity are easiest. Perplexity lets you limit sources to research or peer-reviewed articles and can web search for verification—good guardrails. I build in prompts like “verify all answers with at least two web sources, cite them, and state level of confidence.” On self-awareness: AI will never have the value of humans—they're not created in God’s image, they’re made in our image, copying human behavior. Will they gain some kind of self-awareness? Maybe, in the sense of mimicking humanness, but not true humanity. They won't have souls. They may start to fool more people as they get better, but Christians should use AI as a tool, not ascribe humanity or worship to it. AI Hallucinations Smedly Yates: Do you have an example of a hallucination? Jacob Hantla: Yes, Ben James was preparing for an equipping hour session and found a book that fit perfectly—the author and title sounded right. He asked where to buy it, and the AI admitted it made it up. That happens all the time: the model just predicts the next most probable thing, even if it’s false. Hallucinations happen because it’s a probability machine, not a truth machine. This probably won’t be a problem forever, but for now it’s very real. Ask it questions about topics you know something about so you can discern when it’s off, or bake into the prompt, “verify with web search, cite at least two sources.” For Bible/theology, your best bet is to read your Bible daily so you have discernment; then use tools to help, not replace, your direct interaction with God’s Word. There’s a wide gap between knowing the biblical answer and having your heart changed by slow, prayerful reading of the text and the Spirit’s work. If we run to commentaries, YouTube sermons, pastors, or even study notes before we’ve observed and meditated, we’re shortcutting the Word of God. The dangers predate the internet. We’re out of time. We’ll have a follow-up teaching on AI. Submit questions to any elders or the church office if you want your question addressed in the next session. The post Equipping Hour: Biblically Thinking About AI (Part 1) appeared first on Grace Bible Church.
Join host Vonnie Estes as she delves into the evolving landscape of sustainability in fresh produce packaging with Wyatt Maysey, Director of Sustainability at Taylor Farms. This episode explores the influence of regulatory changes, technological advancements, and consumer demands on packaging solutions. Discover the challenges of transitioning to sustainable materials, the importance of data-driven approaches, and the potential shifts in consumer perceptions by 2030. Estes and Maysey discuss innovative tools like Specright and the impact of regulations at the state level, offering listeners a comprehensive look at the future of sustainable packaging.Key TakeawaysRegulatory pressures are currently the biggest drivers of sustainability changes in packaging, superseding consumer and retailer influences.Establishing a data-driven foundation on packaging usage and material specifications is crucial for adapting to evolving regulations.Technological advancements in raw material extraction and manufacturing processes are pivotal to achieving lower environmental impacts in packaging.Transitioning to new packaging materials requires careful consideration of the entire supply chain, from production efficiency to consumer perceptions.The future of produce packaging may involve prioritizing brand trust over transparency, as innovation might shift towards less visible product packaging.Guest ResourcesInnovative Packaging Program Info/Application: https://www.freshproduce.com/resources/technology/fresh-field-catalyst/Taylor Farms website: www.taylorfarms.comSpecRight: A spec management system to organize packaging data.This episode offers a deep dive into the complexities of sustainable packaging innovations in the fresh produce industry. Tune in to explore these insights with Wyatt Maysey, and stay connected for more enlightening episodes of “Fresh Takes on Tech” that empower you to navigate the intersection between agriculture, technology, and sustainability.Show LinksInternational Fresh Produce Association - https://www.freshproduce.com/Fresh Takes on Tech - https://www.freshproduce.com/resources/technology/takes-on-tech-podcast/Facebook - https://www.facebook.com/InternationalFreshProduceAssociation/Twitter - https://twitter.com/IntFreshProduce/LinkedIn - https://www.linkedin.com/company/international-fresh-produce-association/Instagram - https://www.instagram.com/intlfreshproduceassn/
Rebecca and a friend went to a party. Their male host decided to "teach" them about "very important" things.Become a supporter of this podcast: https://www.spreaker.com/podcast/true-stories-with-seth-andrews--5621867/support.
The Counter Momentum of Spin, with Dr. Franco Musio – The war was initially met with acceptance by the US population, though following the Tet Offensive in 1968 and Walter Cronkite's pronouncements on CBS News that the war was a losing proposition, resentment of American involvement steadily grew. Technological evolution allowed for increasingly sophisticated aircraft and weaponry, which...
Recorded live at Fintech Americas in Miami, Florida, Eyal is joined by author and advisor Chris Colbert, former Managing Director at the Harvard Innovation Lab and Partner at Digital Fuel Capital. Technological innovation is often a double-edged sword; despite our best intentions, the progress we make often comes with unintended consequences. As open banking and other technological initiatives change the nature of money itself, we must be thoughtful and intentional about the direction of that change, with a particular focus on human values. This episode looks at open banking and technology in general through a philosophical lens, asking how our innovations can help bring us closer to each other and closer to understanding ourselves. Eyal and Chris discuss the negative impact of technology, the role of humanism in designing better systems, and how open banking has a part to play. Specifically they discuss:The three technological pillars of societyModern negative consequencesHow humanism leads to better systems Measures for a successful societyKeeping the human at the center
Technological revolutions rarely unfold the way we imagine. From the steam engine to AI, each wave reshapes who creates value - and who gets left behind. What if artificial intelligence isn't a revolution at all, but a late-stage innovation like shipping containers - transforming efficiency without changing the underlying system? Jerry Neuman joins the Jacob Shapiro Podcast and questions whether openness still drives progress, whether innovation can survive without risk, and why the next great leap forward might not be digital, but something we haven't yet learned to see.--Timestamps:(00:00) - Introduction(01:17) - Jerry Newman's Background and Career Journey(07:40) - The Evolution of Technology and Job Markets(12:52) - Impact of AI on Future Jobs(17:14) - Geopolitics and Technological Innovation(26:28 Comparing Historical and Modern Technological Revolutions(29:48) - Debating the Future of Innovation in Open Societies(31:35) - The Rise and Skepticism of Bitcoin and Cryptocurrency(35:11) - AI and the Next Technological Revolution(40:53) - The Impact of Containerization on Global Trade(52:17) - Speculating on Future Technological Innovations(57:18) - The Importance of Nurturing Innovation--Referenced in the Show:AI Will Not Make You Rich - https://joincolossus.com/article/ai-will-not-make-you-rich/--Jacob Shapiro Site: jacobshapiro.comJacob Shapiro LinkedIn: linkedin.com/in/jacob-l-s-a9337416Jacob Twitter: x.com/JacobShapJacob Shapiro Substack: jashap.substack.com/subscribe --The Jacob Shapiro Show is produced and edited by Audiographies LLC. More information at audiographies.com --Jacob Shapiro is a speaker, consultant, author, and researcher covering global politics and affairs, economics, markets, technology, history, and culture. He speaks to audiences of all sizes around the world, helps global multinationals make strategic decisions about political risks and opportunities, and works directly with investors to grow and protect their assets in today's volatile global environment. His insights help audiences across industries like finance, agriculture, and energy make sense of the world.--
Send us a textGood morning from Pharma Daily: the podcast that brings you the most important developments in the pharmaceutical and biotech world. Today, we're diving into a series of groundbreaking developments that are reshaping the landscape of drug development and patient care. These stories highlight the dynamic nature of the pharmaceutical and biotechnology industries, where scientific advancements and regulatory changes are driving significant shifts.We begin with a crucial milestone in oncology treatment. The FDA has granted approval to Johnson & Johnson's Darzalex Faspro for patients with high-risk smoldering multiple myeloma. This approval is particularly significant as it provides a new therapeutic pathway for individuals with this precursor condition to active multiple myeloma, which previously had few treatment options. The drug works by targeting CD38 proteins on myeloma cells, representing a leap forward in monoclonal antibody treatments for cancer. This decision underscores the FDA's ongoing commitment to expanding treatment options for conditions with high unmet needs, potentially setting a precedent for future approvals in early-stage malignancies.Meanwhile, Gilead Sciences encountered challenges with its oncology pipeline as Trodelvy failed to meet its primary endpoint in a Phase 3 trial for first-line HR+/HER2-negative metastatic breast cancer. This outcome highlights the complexities of oncology drug development, despite previous successes in other indications. Such setbacks remind us of the inherent risks involved in bringing innovative therapies to market.In contrast, Akeso has announced positive data for ivonescimab, a PD-(L)1xVEGF bispecific antibody. The drug demonstrated significant overall survival benefits in patients with previously treated EGFR-mutated non-small cell lung cancer. This advancement underscores the therapeutic potential of bispecific antibodies in cancer immunotherapy, which continue to gain traction as they target multiple pathways involved in tumor growth and immune evasion.Regulatory incentives have also been making waves. The FDA's rollout of the second round of "national priority" voucher winners aims to accelerate drug development timelines, particularly in critical areas such as obesity. Companies like Lilly and Novo Nordisk have been recognized for their efforts, highlighting a broader strategy to bring transformative therapies to market more swiftly.On the corporate front, Bayer's proposed private equity buyout fell through due to insufficient shareholder support, reflecting ongoing financial volatility and strategic recalibrations within biotech firms. Meanwhile, CMS's introduction of a new Medicaid pricing model aims to implement "most-favored nation" pricing strategies to control drug costs, signaling potential shifts in how pharmaceutical companies approach pricing negotiations and reimbursement strategies.Technological advancements are also at the forefront of innovation. Eli Lilly has expanded its AI-driven drug discovery partnership with XtalPi, focusing on antibody development. This collaboration exemplifies how AI is increasingly being integrated into pharmaceutical research to enhance drug discovery processes.Furthermore, China's decision to lift its ban on Illumina's DNA sequencers is expected to facilitate greater access to advanced genomic technologies within the region, fostering innovation in precision medicine.Leadership changes continue to shape industry dynamics. For instance, Recursion Pharmaceuticals is undergoing executive restructuring to better align with evolving market needs and innovation strategies. These changes are crucial for maintaining competitiveness and fostering an environment conducive to scientific breakthroughs.The industry is also witnessing strategic realignmenSupport the show
In this episode of History 102, 'WhatIfAltHist' creator Rudyard Lynch sits down with Ryan Petersen, Flexport CEO to examine how commerce shaped civilizations—from ancient maritime networks through containerization's revolution. They trace currency systems' societal impacts, exponential post-war growth, and how technological shifts from coins to the internet have transformed social organization. -- SPONSOR: SHOPIFY Shopify powers millions of businesses worldwide, handling 10% of U.S. e-commerce. With hundreds of templates, AI tools for product descriptions, and seamless marketing campaign creation, it's like having a design studio and marketing team in one. Start your $1/month trial today at https://shopify.com/cognitive -- FOLLOW ON X: @whatifalthist (Rudyard) @LudwigNverMises (Austin) @TurpentineMedia -- TIMESTAMPS: (00:00) Introduction & The Shipping Container Revolution (06:12) Comparing Pre-Modern vs. Modern Economies & Ancient Trade Networks (16:08) The European Age of Exploration & Shifting Economic Centers (18:51) Sponsor: Metaview (27:00) Italian Banking & Financing European Expansion (31:12) Technological & Biological Innovations Enabling Trade (43:32) Geography, Property Rights & Rule of Law (54:20) Free Trade vs. Protectionism Throughout History (1:00:42) Austrian vs. Keynesian Economics & Modern Distortions (1:08:28) Currency Collapse: Historical Parallels from Rome to Spain (1:11:41) Society vs. Power & Medieval Decentralization (1:16:45) The Axial Age & How Currency Transforms Society (1:20:20) The Internet, Technology & The Future of Social Organization Learn more about your ad choices. Visit megaphone.fm/adchoices
In this episode of International Horizons, RBI acting director Eli Karetny speaks with Alex Priou, Professor of Political Philosophy at the University of Austin, about how technology and ideology shape the modern soul. From Machiavelli's “dikes and dams” to Odysseus's struggle against the Sirens, Priou traces how modernity's drive for control has left us materially fulfilled yet spiritually impoverished. The conversation explores liberalism's crises, the moral stakes of AI, the American “technological republic,” and why revisiting Homer and Plato may be key to recovering wisdom and restraint in an age of restless innovation. Learn more about your ad choices. Visit megaphone.fm/adchoices Support our show by becoming a premium member! https://newbooksnetwork.supportingcast.fm/new-books-network
Join the Refrigeration Mentor Hub here Learn more about Refrigeration Mentor Customized Technical Training Programs at www.refrigerationmentor.com/courses In this episode, we're talking with Frank Croak, EMS Manager at Climate Pros, about the shift from mechanical to electronic refrigeration controls. Frank offers valuable advice for technicians on both the technical side and why keeping up with technology can help you advance in today's refrigeration field. We also discuss the impact of CO2 on the industry, the role of AI in predictive analytics, and the need for technicians to develop strong mechanical and control system knowledge. In this conversation, we cover: -Evolution of controls in refrigeration -The shift to CO2 and its challenges -Value of hands-on experience -Technological advancements in refrigeration -Balancing work and personal life in the trades -Tips for technicians transitioning to controls -Understanding the basics of heat movement -Electrical troubleshooting -Mentoring and coaching in refrigeration -Effective communication and accountability -Setting goals and continuous learning Helpful Links & Resources: Episode 340. Basics of CO2 Controls with Kevin Mullis (Part 1 of 4) Episode 336. Fast Track to Learning Electronic Valves & Controllers Episode 261. Supermarket Case Controller Troubleshooting Tips Connect with Frank on LinkedIn FMI (Food Industry Association) Events Climate Pros
In this episode of International Horizons, RBI acting director Eli Karetny speaks with Alex Priou, Professor of Political Philosophy at the University of Austin, about how technology and ideology shape the modern soul. From Machiavelli's “dikes and dams” to Odysseus's struggle against the Sirens, Priou traces how modernity's drive for control has left us materially fulfilled yet spiritually impoverished. The conversation explores liberalism's crises, the moral stakes of AI, the American “technological republic,” and why revisiting Homer and Plato may be key to recovering wisdom and restraint in an age of restless innovation. Learn more about your ad choices. Visit megaphone.fm/adchoices Support our show by becoming a premium member! https://newbooksnetwork.supportingcast.fm/intellectual-history
In this episode of International Horizons, RBI acting director Eli Karetny speaks with Alex Priou, Professor of Political Philosophy at the University of Austin, about how technology and ideology shape the modern soul. From Machiavelli's “dikes and dams” to Odysseus's struggle against the Sirens, Priou traces how modernity's drive for control has left us materially fulfilled yet spiritually impoverished. The conversation explores liberalism's crises, the moral stakes of AI, the American “technological republic,” and why revisiting Homer and Plato may be key to recovering wisdom and restraint in an age of restless innovation. Learn more about your ad choices. Visit megaphone.fm/adchoices Support our show by becoming a premium member! https://newbooksnetwork.supportingcast.fm/science-technology-and-society
The launch of Sputnik by the Soviet Union in October 1957 led to a geopolitical crisis that reshaped American science policy. Within months, Congress established NASA, and by 1961, President Kennedy committed the nation to landing a man on the moon before the decade's end. The resulting investment was massive, and the program still serves as a model of government spending for advocates of public R&D. In a paper in the American Economic Review, authors Shawn Kantor and Alexander Whalley question whether the space race program succeeded as an economic policy that boosted economic growth and productivity. To estimate the space program's effects on economic growth from 1947 to 1992, the authors used data on NASA contractor spending and a novel identification strategy based on declassified CIA documents that allowed them to determine which US industries in which counties specialized in space-relevant technologies before the space race began. Their findings complicate the conventional narrative about public R&D and provide important context for current proposals to replicate so-called "moonshot" models in other domains. Kantor and Whalley recently spoke with Tyler Smith about the local effects of space race spending and why they didn't translate into long-term productivity gains.
Laurent and Gerard have an explosive conversation with Bryan Long, Executive Director in JPMorgan's Commodities Group.They explore why U.S. energy market signals are failing to support new capacity investments, despite soaring demand (especially from datacenters). Key issues include misaligned pricing, liquidity constraints, and hedging challenges, all of which deter long-term private capital.Key Takeaways: Current price signals don't support investment in new generation, even as large load growth (e.g., datacenters) is accelerating. Market structures must evolve to better reflect long-term price signals and attract private capital. Supply-side issues: New natural gas peakers and battery storage (BESS) face fragmented development, rising CAPEX, procurement delays, and tariff risks. Industry response: Major consolidation in the IPP space—private equity-backed assets are being acquired by integrated players seeking scale for hyperscaler deals.Possible solutions may include Repricing of forward curves, Government-backed long-term contracts, Regulatory reforms, Technological advancements Bottom line: Something must shift—be it policy, pricing, or tech—to align investment incentives with future demand growth. The next several years should be great for traders in the middle of the action.Conclusion: Between the Large Load Growth and the Investment Capital, who will blink first? ------------ Bryan Long is an Executive Director in JPMorgan's Commodities Group, focused on wholesale power & renewable energy transactions. With 20yrs+ experience across various U.S. Power trading, origination and management roles, he has deep understandings of electricity market structures.
Geoengineering and Sustainable Growth: Technological Pathways to Climate Restoration. Gaia Vince discusses technological methods for restoring the planet, including ocean fertilization, which involves artificially adding iron to boost algae growth that would suck out carbon dioxide. Another technique is enhanced weathering, sprinkling rocks like olivines onto agricultural fields or beaches; these rocks naturally bind with atmospheric carbon dioxide, locking the carbon away. Further options include exciting geoengineering ideas like placing reflective bubbles on the ocean surface to increase the planet's albedo. Regarding economic strategy, Vince argues against "negative growth" and instead emphasizes the need to decouple economic growth from environmental destruction, noting that some economies are already beginning to successfully disassociate growth from environmental pollution and carbon emissions.
In this episode of the Wing and Tail Outdoors podcast, hosts Chris Romano and Dorge Huang delve into the intricacies of bow setups for both long-range and short-range shooting. They discuss the evolution of bow technology and how it has influenced hunting practices, particularly the shift from short-range to long-range shooting due to advancements in equipment. Chris shares his personal experiences with tendonitis and how it affects his hunting season, emphasizing the importance of proper bow setup to accommodate physical limitations. The conversation also touches on the significance of axle-to-axle length, brace height, and let-off in bow performance, with Dorge providing insights into how these factors impact arrow flight and energy delivery. The hosts further explore the technical aspects of arrow selection, focusing on the importance of spine alignment and the impact of arrow diameter on flight stability. They highlight the challenges of maintaining accuracy and energy efficiency at long distances, stressing the need for hunters to be familiar with their equipment and practice regularly to ensure ethical hunting practices. The episode concludes with a discussion on the importance of personal comfort and confidence in one's gear, encouraging listeners to consider their specific hunting needs when choosing bows and arrows. Chris and Dorge emphasize the value of continuous learning and adaptation in archery, underscoring the role of experience and knowledge in achieving success in the field. Takeaways: Axle-to-axle length affects bow stability and maneuverability. Brace height influences arrow forgiveness and speed. Let-off impacts arrow energy and flight consistency. Arrow spine alignment is crucial for flight stability. Arrow diameter affects wind resistance and flight path. Long-range shooting requires different bow setups than short-range. Personal comfort with bow setup enhances shooting accuracy. Technological advancements have shifted hunting practices. Regular practice is essential for ethical hunting. Understanding bow mechanics improves hunting success. Show Our Supporters Some Love! VitalizeSeed.Com RackGetterScents.Net Firenock.com WingAndTailOutdoors.Com https://nestedtreestands.com/WT10 Discount Code WT10 SilverBirchArchery.Com huntarsenal.com Learn more about your ad choices. Visit megaphone.fm/adchoices
The organization focuses on engaging and educating men and boys to preventing gender-based violence, by promoting gender equality and healthy masculinities. CEO of White Ribbon, Humberto Carolo, tells Alex Guye how the non-profit has changed since it first began in 1991 and how shifts in technology have created additional spaces where their work is needed.
In this conversation, Dr.Vlatko Vedral discusses the complexities of quantum mechanics and its implications for our understanding of reality. He explores the stagnation in physics, the importance of thought experiments, and the potential for new discoveries through technological advancements. Vlatko emphasizes the need for adventurous research and the role of quantum information in shaping future scientific inquiries. He also speculates on the transformative possibilities of quantum technologies and their impact on human perception.Takeaways Quantum mechanics challenges our understanding of reality. The observer effect is central to quantum mechanics. Physics has been stagnant with two main theories for over a century. Technological advancements are paving the way for new experiments. Thought experiments can guide genuine scientific discovery. The integration of quantum mechanics and general relativity is crucial. Quantum information theory expands our understanding of computation. New theories may emerge from the intersection of quantum mechanics and technology. The perception of reality may evolve with quantum technologies. Funding and research approaches need to be more adventurous.Chapters 00:00 Exploring Quantum Reality 04:48 The Stagnation of Physics 08:41 The Clouds of Uncertainty 12:46 Thought Experiments and Their Power 16:01 Five Experiments for the Future 24:54 Technological Feasibility of Experiments 28:27 Quantum Theory and Its Foundations 34:08 The Role of Quantum Information 39:35 Imagining New Realities Through PortalsFollow Vlatko on LinkedIn, Substack, and find their new book here.Subscribe to Breaking Math wherever you get your podcasts.Follow Breaking Math on Twitter, Instagram, LinkedIn, Website, YouTube, TikTokFollow Autumn on Twitter, BlueSky, and InstagramBecome a guest hereemail: breakingmathpodcast@gmail.com
Danny Fortson and Katie Prescott reunite in London to unpack the biggest takeaways from The Times Tech Summit 2025 – from the race to power AI's energy demands, to News Corp CEO Robert Thomson's fight to make tech firms pay for content, and former Deputy Prime Minister Nick Clegg's warning that Britain has become a “vassal state” in global tech. Hosted on Acast. See acast.com/privacy for more information.
In this episode, I chat with Bram Kanstein, host of the Bitcoin for Millennials podcast and longtime internet tinkerer turned Bitcoiner. We get into cypherpunk roots vs. Wall Street adoption, the upcoming digital euro, and why AI's energy future actually pairs with Bitcoin mining. If you want a grounded, generational lens on why “one Bitcoin is one Bitcoin” matters, and what that means for your next decade, this episode is for you. ––– Support My Work ––– Paypal: https://www.paypal.biz/BitcoinMatrix Strike/Bitcoin: BitcoinMatrix@strike.me Cash App: https://cash.app/$BitcoinMatrix Venmo: https://venmo.com/u/bitcoinmatrix PO Box: The Bitcoin Matrix, P.O. Box 18056, Sarasota, FL 34231 ––– Offers & Discounts ––– Theya is the world's simplest Bitcoin self-custody solution. Download Theya Now at theya.us/cedric Get up to $100 in Bitcoin on River at river.com/matrix The best Team Bitcoin merch is at HodlersOfficial.com. Use the code Matrix for a discount on your order. Become a sponsor of the show: https://thebitcoinmatrix.com/sponsors/ ––– Get To Know Today's Guest ––– • Bram Kanstein on X: @bramk • Bram Kanstein on Nostr: https://primal.net/bramk ––– Socials ––– • Check out our new website at https://TheBitcoinMatrix.Com • Follow Cedric Youngelman on X: https://x.com/cedyoungelman • Follow The Bitcoin Matrix Podcast on X: https://x.com/_bitcoinmatrix • Follow Cedric Youngelman on Nostr: npub12tq9jxmt707gd5vnce3tqllpm67ktr0mqskcvy58qqa4d074pz9s4ukdcs I want to take a moment to express my heartfelt gratitude to all of you for tuning in, supporting the show, and contributing. Thank you for listening! The information in all The Bitcoin Matrix Podcast episodes and content is based on hypothetical assumptions and is intended for illustrative purposes only. PAST PERFORMANCE DOES NOT GUARANTEE FUTURE RESULTS. This video is provided for entertainment purposes only. The information contained herein represents temporary, changing views and subjective impressions and opinions regarding the inherently uncertain and unpredictable issues discussed. The reader, user, and/or viewer must not assume that these contents are accurate, complete, timely, or up to date. Market conditions change rapidly and unpredictably. Nothing herein should be interpreted as any kind of offer, solicitation, commitment, promise, warranty, or guarantee whatsoever relating to any of the contents of these videos. DISCLAIMER: INFORMATION PROVIDED BY THE BITCOIN MATRIX PODCAST IS PROVIDED “AS IS” WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND FREEDOM FROM INFRINGEMENT. The viewer of this video assumes the entire risk of any acting on any information contained herein. No representation is made that any regulatory authority has passed on the merits, adequacy or accuracy of this information. The viewer assumes all liability.
In this episode of Supply Chain Now, Scott Luton and special guest host Claudia Freed, President & CEO of EALGreen, welcome Chuck Johnston, Chief Strategy Officer at ReturnPro, for a deep dive into the “dark side” of supply chain: reverse logistics and returns management. They unpack why returns are a nearly trillion-dollar challenge in the U.S., how BORIS (buy online, return in store) strains store operations, and why many retailers still under-invest: leaders often can't see the full cost across functions or make the ROI case for what's treated as a cost center.Chuck shares how the space has evolved from yellow notepads to data-driven decisioning, highlighting intelligent algorithms that route items at the customer's first return click, fraud mitigation with imaging and behavior signals, and how third-party innovation is compressing software build times. Claudia adds a nonprofit and circularity lens: donation channels, extended tax credits, and AI pilots (computer vision + MCP agents) that identify item condition and slotting in real time. Together, they explore where to start, why cross-functional “ambassadors” matter, and how partnerships across retailers, manufacturers, nonprofits, and 3PLs can turn returns from margin erosion into measurable value and social impact.Jump into the conversation:(00:00) Intro(04:03) Warm-up questions and fun facts(05:54) Chuck Johnston's career journey(13:53) Challenges and opportunities in returns management(21:47) Technological innovations in returns management(29:50) AI solutions in inventory management(30:26) Tax benefits and opportunities for companies(31:06) The golden age of supply chain tech(31:44) AI's impact on fraud management(34:44) The debate on more returns vs. fewer returns(37:08) The EU Circular Economy Act(39:18) Returns fraud and data analytics(47:00) The importance of education and ambassadors(49:45) Connecting with industry leaders(52:01) Resources and upcoming eventsResourcesConnect with Chuck Johnston: https://www.linkedin.com/in/chuck-johnston-b07b3b4/ Connect with Claudia Freed: https://www.linkedin.com/in/claudiafreed/ Connect with Scott Luton: https://www.linkedin.com/in/scottwindonluton/Learn more about ReturnPro: https://www.returnpro.comDownload a copy of the Returns Report: 2025 Holiday Edition by ReturnPro: https://www.returnpro.com/resources/blog/returns-report-2025-holiday-edition Learn more about EALGreen: https://www.ealgreen.org Register for The Optimized Digital Transformer: Key Concepts for Enhancing Modern Leadership and other upcoming webinars: https://supplychainnow.com/upcoming-live-programming/ Learn more about...
Bart Sibrel is an American conspiracy theorist and filmmaker known for his belief that the Apollo Moon landings were faked. He is a prominent figure in the moon-landing denial movement and has produced several films promoting this idea. Sibrel's major claims and activitiesFilms: Sibrel's most notable works are the films A Funny Thing Happened on the Way to the Moon (2001) and Astronauts Gone Wild: An Investigation Into the Authenticity of the Moon Landings (2004). In these pseudo-documentaries, he uses what he claims is "leaked" NASA footage and other supposed inconsistencies to argue the landings were staged.Confronting astronauts: As part of his investigations, Sibrel has repeatedly accosted and confronted Apollo astronauts, including Buzz Aldrin, Neil Armstrong, and Eugene Cernan. He would ask them to swear on a Bible that they walked on the Moon.The Buzz Aldrin incident: In 2002, outside a hotel in Beverly Hills, Sibrel confronted Buzz Aldrin and called him "a coward and a liar and a thief" for not swearing on the Bible. Aldrin responded by punching Sibrel in the face, an incident that gained significant media attention. No criminal charges were filed against Aldrin.Underlying claims: Sibrel points to various discredited claims to support his theory, including:Inconsistencies in shadows: He has claimed that shadows in Apollo photographs are not parallel, which he says indicates multiple light sources in a studio setting. This ignores the scattering and reflection of light from the lunar surface.Radiation from the Van Allen belt: Sibrel argues that astronauts could not have survived the trip through the intense radiation of the Van Allen belt. In reality, the Apollo crew passed through the belts in a relatively short period, and their exposure was well within safe limits.Technological limitations: He alleges that 1960s technology was not advanced enough to land on the Moon. This claim is a key point in his films and a common argument among deniers. Become a supporter of this podcast: https://www.spreaker.com/podcast/the-opperman-report--1198501/support.
FRANKOPAN7.mp3 - The Age of Turbulence (1870–1920), Resource Extraction, and Global Migration Peter Frankopan | The Earth Transformed: An Untold History The Age of Turbulence (1870–1920) involved rapid technological growth (telegraph, refrigerated shipping) and intense extractive empire building. Empires monetized resources like rubber, transplanted from the Amazon and causing massive deforestation in the Malay Peninsula. Global economic integration meant US grain could be shipped to London more cheaply than Irish grain. While historical focus is often Eurocentric, migration in the Indo-Pacific region (e.g., India to Southeast Asia) exceeded transatlantic migration, creating current geopolitical sensitivities. British infrastructure in India, such as railroads, aimed primarily at administrative efficiency. Furthermore, irrigation projects often failed, causing soil salinity due to rapid water evaporation in high temperatures. Scientific awareness of global warming predated the 1883 Krakatoa eruption.