Podcasts about V3

  • 437PODCASTS
  • 925EPISODES
  • 47mAVG DURATION
  • 5WEEKLY NEW EPISODES
  • Feb 6, 2026LATEST

POPULARITY

20192020202120222023202420252026

Categories



Best podcasts about V3

Show all podcasts related to v3

Latest podcast episodes about V3

SynGAP10 weekly 10 minute updates on SYNGAP1 (video)
All #SYNGAP1 Families need to take part in our Natural History Studies: ProMMiS & Citizen #S10e198

SynGAP10 weekly 10 minute updates on SYNGAP1 (video)

Play Episode Listen Later Feb 6, 2026 9:58


Thursday, February 5, 2026 - Week 6 Happy #RareDisease & #BlackHistory Month!   #NaturalHistory means how this disease progresses.  Reminder: We have only been at this for 17 years, first patients were identified via Hamdan, 2009. https://pubmed.ncbi.nlm.nih.gov/19196676/   Retrospective Digital NHS: cureSYNGAP1.org/Citizen (Growing list of tools available to families, for free)   Prospective Multi-disciplinary Multi-site NHS: ProMMiS cureSYNGAP1.org/ProMMiS   Reminder, only possible by CS1 support for non-CHOP sites and travel plus huge gift to Penn. https://www.chop.edu/news/25-million-gift-penn-medicine-and-children-s-hospital-philadelphia-establishes-center-epilepsy   Potential for being a control arm in the future.   Protocol: https://www.linkedin.com/posts/curesyngap1_syngap1-stxbp1-dee-activity-7425223573134327808-SVEQ & early data: https://pubmed.ncbi.nlm.nih.gov/40119723/   Join the ~160 families who have enjoyed excellent clinical care and contributed tot he future of SYNGAP1.  Today, a 4 month old is going! CHOP: 119 new, V2- 67, V3- 32, V4- 10, V5- 4 CHCO: 37 new, V2- 7 Stanford: 8 new, V2- 2 Total: 164 (double counting one family who goes to multiple sites)   Survey English: https://curesyngap1.org/SurveyProMMiS Spanish: https://curesyngap1.org/encuestaProMMiS   94 Responses to survey, so far: Why not? Did not receive an invitation, Too far to travel, Too expensive Barriers: Logistics, Cost, Time off, Behaviors, Insurance   ETC. Pubmed 2026 is at 6!  But will soon be 7 with the McKee paper! https://pubmed.ncbi.nlm.nih.gov/?term=syngap1&filter=years.2026-2026&sort=date   Biorepository needs more samples.  Check out the list and map here https://docs.google.com/presentation/d/1IjaHILXj7AlBDlbTJgvYrkBS_0bnI8VCnTIiPXJ7JGM/edit?usp=sharing and contribute blood.  The data and research we do with these samples is invaluable.   May 28, San Francisco, CA: cureSYNGAP1.org/SF26   SOCIAL MATTERS 4,668 LinkedIn.  https://www.linkedin.com/company/curesyngap1/ 1,520 YouTube.  https://www.youtube.com/@CureSYNGAP1 11.2k Twitter https://twitter.com/cureSYNGAP1 45k Insta https://www.instagram.com/curesyngap1/   $CAMP stock is at $3.59 on 5 Feb. ‘26 https://www.google.com/finance/beta/quote/CAMP:NASDAQ   Like and subscribe to this podcast wherever you listen.  https://curesyngap1.org/podcasts/syngap10/ Episode 198 of #Syngap10 #CureSYNGAP1 #Podcast

Doubles Only Tennis Podcast
Tennis Gear Deep Dive: Dampeners, Grips, Bags, & Training with ADV Founder, Lavie Sak

Doubles Only Tennis Podcast

Play Episode Listen Later Jan 26, 2026 39:24


This episode is a little different. I'm diving into tennis gear, everything except the racquet and shoes, with the founder of one of my favorite brands, ADV. I use their bags, dampeners, and even sweat bands.Lavie Sak has a tech background, plays tennis, and used to coach as well. We explore how his company lets players lead the design of its products. From dampeners to grips to their popular bags, ADV innovates as well as any company in tennis. Lavie shares the messy first prototype, the tough cuts, and why ADV chose quality over mass pricing while partnering with pros who give real feedback.How ADV got startedDampener testing across 27 racquets and sound profilesTennis grips - how to choose between dry and tackyDesigning the ADV Pro bag and prioritizing featuresHow they develop an idea into a finished productFeedback loops that shaped V2 and V3 of the bagWhy ADV makes two backpacksCurated training kit components and use casesPricing tradeoffs, materials, and longevityDoubles tips on serve variety and aiming middlePartnerships with Sem Verbeek, JP Smith, and Zus TennisI use the ADV Pro for travel and the Flex bag locally around Fort Worth.Links:Shop ADV TennisLearn more about ADV & follow:ADV Tennis - InstagramADV Tennis - YouTubeADV Tennis - Facebook ----- **Join the #1 Doubles Strategy Newsletter for Club Tennis Players** New doubles strategy lessons weekly straight to your inbox **Become a Tennis Tribe Member**Tennis Tribe Members get access to premium video lessons, a monthly member-only webinar, doubles strategy Ebooks & Courses, exclusive discounts on tennis gear, and more. Learn More & Sign Up Here **Other Free Doubles Content** Serve Strategy Cheatsheet Return Strategy Cheatsheet Serve Strategy 101 - Video Course

Embedded
519: The Password Is All Zeros

Embedded

Play Episode Listen Later Jan 23, 2026 66:51


Mark Omo and James Rowley spoke with us about safecracking, security, and the ethics of doing a bad job. Mark and James gave an excellent talk on the development of their safecracking tools at DEF CON 33: Cash, Drugs, and Guns: Why Your Safes Aren't Safe. It included a section of interaction involving the lock maker's lawyers bullying them and how the Electronic Frontier Foundation (EFF) has a Coders' Rights Project to support security research. As mentioned in the show, the US Cyber Trust Mark baseline has a very straightforward checklist; NISTIR 8259 is the overall standard, NISTIR 8259A is the technical checklist, NISTIR 8259B is the non-technical (process/maintenance) checklist. Roughly the process is NISTIR 8259 -> Plan/Guidance; NISTIR 8259A -> Build; NISTIR 8259B -> Support. We discussed ETSI EN 303 645 V3.1.3 (2024-09) Cyber Security for Consumer Internet of Things: Baseline Requirement and the EU's CRA: Cyber Resilience Act which requires manufacturers to implement security by design, have security by default, provide free security updates, and protect confidentiality. See more here: How to prepare for the Cyber Resilience Act (CRA): A guide for manufacturers. We didn't mention Ghidra in the show specifically, but it is a tool for reverse engineering software: given a binary image, what was the code? Some of the safecracking was helped by the lock maker using the same processor in the PS4 which has many people looking to crack it. See fail0verflow :: PS4 Aux Hax 1: Intro & Aeolia for an introduction.  Mark and James have presented multiple times at Hardwear.io, a series of conferences and webinars about security (not wearables). Some related highlights: 2024: Breaking Into Chips By Reading The Datasheet is about the exploit developed for the older lock version on the safes discussed in the show. USA 2025: Extracting Protected Flash With STM32-TraceRip is about STM32 exploits.

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0
Artificial Analysis: Independent LLM Evals as a Service — with George Cameron and Micah-Hill Smith

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

Play Episode Listen Later Jan 8, 2026 78:24


Happy New Year! You may have noticed that in 2025 we had moved toward YouTube as our primary podcasting platform. As we'll explain in the next State of Latent Space post, we'll be doubling down on Substack again and improving the experience for the over 100,000 of you who look out for our emails and website updates!We first mentioned Artificial Analysis in 2024, when it was still a side project in a Sydney basement. They then were one of the few Nat Friedman and Daniel Gross' AIGrant companies to raise a full seed round from them and have now become the independent gold standard for AI benchmarking—trusted by developers, enterprises, and every major lab to navigate the exploding landscape of models, providers, and capabilities.We have chatted with both Clementine Fourrier of HuggingFace's OpenLLM Leaderboard and (the freshly valued at $1.7B) Anastasios Angelopoulos of LMArena on their approaches to LLM evals and trendspotting, but Artificial Analysis have staked out an enduring and important place in the toolkit of the modern AI Engineer by doing the best job of independently running the most comprehensive set of evals across the widest range of open and closed models, and charting their progress for broad industry analyst use.George Cameron and Micah-Hill Smith have spent two years building Artificial Analysis into the platform that answers the questions no one else will: Which model is actually best for your use case? What are the real speed-cost trade-offs? And how open is “open” really?We discuss:* The origin story: built as a side project in 2023 while Micah was building a legal AI assistant, launched publicly in January 2024, and went viral after Swyx's retweet* Why they run evals themselves: labs prompt models differently, cherry-pick chain-of-thought examples (Google Gemini 1.0 Ultra used 32-shot prompts to beat GPT-4 on MMLU), and self-report inflated numbers* The mystery shopper policy: they register accounts not on their own domain and run intelligence + performance benchmarks incognito to prevent labs from serving different models on private endpoints* How they make money: enterprise benchmarking insights subscription (standardized reports on model deployment, serverless vs. managed vs. leasing chips) and private custom benchmarking for AI companies (no one pays to be on the public leaderboard)* The Intelligence Index (V3): synthesizes 10 eval datasets (MMLU, GPQA, agentic benchmarks, long-context reasoning) into a single score, with 95% confidence intervals via repeated runs* Omissions Index (hallucination rate): scores models from -100 to +100 (penalizing incorrect answers, rewarding ”I don't know”), and Claude models lead with the lowest hallucination rates despite not always being the smartest* GDP Val AA: their version of OpenAI's GDP-bench (44 white-collar tasks with spreadsheets, PDFs, PowerPoints), run through their Stirrup agent harness (up to 100 turns, code execution, web search, file system), graded by Gemini 3 Pro as an LLM judge (tested extensively, no self-preference bias)* The Openness Index: scores models 0-18 on transparency of pre-training data, post-training data, methodology, training code, and licensing (AI2 OLMo 2 leads, followed by Nous Hermes and NVIDIA Nemotron)* The smiling curve of AI costs: GPT-4-level intelligence is 100-1000x cheaper than at launch (thanks to smaller models like Amazon Nova), but frontier reasoning models in agentic workflows cost more than ever (sparsity, long context, multi-turn agents)* Why sparsity might go way lower than 5%: GPT-4.5 is ~5% active, Gemini models might be ~3%, and Omissions Index accuracy correlates with total parameters (not active), suggesting massive sparse models are the future* Token efficiency vs. turn efficiency: GPT-5 costs more per token but solves Tau-bench in fewer turns (cheaper overall), and models are getting better at using more tokens only when needed (5.1 Codex has tighter token distributions)* V4 of the Intelligence Index coming soon: adding GDP Val AA, Critical Point, hallucination rate, and dropping some saturated benchmarks (human-eval-style coding is now trivial for small models)Links to Artificial Analysis* Website: https://artificialanalysis.ai* George Cameron on X: https://x.com/georgecameron* Micah-Hill Smith on X: https://x.com/micahhsmithFull Episode on YouTubeTimestamps* 00:00 Introduction: Full Circle Moment and Artificial Analysis Origins* 01:19 Business Model: Independence and Revenue Streams* 04:33 Origin Story: From Legal AI to Benchmarking Need* 16:22 AI Grant and Moving to San Francisco* 19:21 Intelligence Index Evolution: From V1 to V3* 11:47 Benchmarking Challenges: Variance, Contamination, and Methodology* 13:52 Mystery Shopper Policy and Maintaining Independence* 28:01 New Benchmarks: Omissions Index for Hallucination Detection* 33:36 Critical Point: Hard Physics Problems and Research-Level Reasoning* 23:01 GDP Val AA: Agentic Benchmark for Real Work Tasks* 50:19 Stirrup Agent Harness: Open Source Agentic Framework* 52:43 Openness Index: Measuring Model Transparency Beyond Licenses* 58:25 The Smiling Curve: Cost Falling While Spend Rising* 1:02:32 Hardware Efficiency: Blackwell Gains and Sparsity Limits* 1:06:23 Reasoning Models and Token Efficiency: The Spectrum Emerges* 1:11:00 Multimodal Benchmarking: Image, Video, and Speech Arenas* 1:15:05 Looking Ahead: Intelligence Index V4 and Future Directions* 1:16:50 Closing: The Insatiable Demand for IntelligenceTranscriptMicah [00:00:06]: This is kind of a full circle moment for us in a way, because the first time artificial analysis got mentioned on a podcast was you and Alessio on Latent Space. Amazing.swyx [00:00:17]: Which was January 2024. I don't even remember doing that, but yeah, it was very influential to me. Yeah, I'm looking at AI News for Jan 17, or Jan 16, 2024. I said, this gem of a models and host comparison site was just launched. And then I put in a few screenshots, and I said, it's an independent third party. It clearly outlines the quality versus throughput trade-off, and it breaks out by model and hosting provider. I did give you s**t for missing fireworks, and how do you have a model benchmarking thing without fireworks? But you had together, you had perplexity, and I think we just started chatting there. Welcome, George and Micah, to Latent Space. I've been following your progress. Congrats on... It's been an amazing year. You guys have really come together to be the presumptive new gardener of AI, right? Which is something that...George [00:01:09]: Yeah, but you can't pay us for better results.swyx [00:01:12]: Yes, exactly.George [00:01:13]: Very important.Micah [00:01:14]: Start off with a spicy take.swyx [00:01:18]: Okay, how do I pay you?Micah [00:01:20]: Let's get right into that.swyx [00:01:21]: How do you make money?Micah [00:01:24]: Well, very happy to talk about that. So it's been a big journey the last couple of years. Artificial analysis is going to be two years old in January 2026. Which is pretty soon now. We first run the website for free, obviously, and give away a ton of data to help developers and companies navigate AI and make decisions about models, providers, technologies across the AI stack for building stuff. We're very committed to doing that and tend to keep doing that. We have, along the way, built a business that is working out pretty sustainably. We've got just over 20 people now and two main customer groups. So we want to be... We want to be who enterprise look to for data and insights on AI, so we want to help them with their decisions about models and technologies for building stuff. And then on the other side, we do private benchmarking for companies throughout the AI stack who build AI stuff. So no one pays to be on the website. We've been very clear about that from the very start because there's no use doing what we do unless it's independent AI benchmarking. Yeah. But turns out a bunch of our stuff can be pretty useful to companies building AI stuff.swyx [00:02:38]: And is it like, I am a Fortune 500, I need advisors on objective analysis, and I call you guys and you pull up a custom report for me, you come into my office and give me a workshop? What kind of engagement is that?George [00:02:53]: So we have a benchmarking and insight subscription, which looks like standardized reports that cover key topics or key challenges enterprises face when looking to understand AI and choose between all the technologies. And so, for instance, one of the report is a model deployment report, how to think about choosing between serverless inference, managed deployment solutions, or leasing chips. And running inference yourself is an example kind of decision that big enterprises face, and it's hard to reason through, like this AI stuff is really new to everybody. And so we try and help with our reports and insight subscription. Companies navigate that. We also do custom private benchmarking. And so that's very different from the public benchmarking that we publicize, and there's no commercial model around that. For private benchmarking, we'll at times create benchmarks, run benchmarks to specs that enterprises want. And we'll also do that sometimes for AI companies who have built things, and we help them understand what they've built with private benchmarking. Yeah. So that's a piece mainly that we've developed through trying to support everybody publicly with our public benchmarks. Yeah.swyx [00:04:09]: Let's talk about TechStack behind that. But okay, I'm going to rewind all the way to when you guys started this project. You were all the way in Sydney? Yeah. Well, Sydney, Australia for me.Micah [00:04:19]: George was an SF, but he's Australian, but he moved here already. Yeah.swyx [00:04:22]: And I remember I had the Zoom call with you. What was the impetus for starting artificial analysis in the first place? You know, you started with public benchmarks. And so let's start there. We'll go to the private benchmark. Yeah.George [00:04:33]: Why don't we even go back a little bit to like why we, you know, thought that it was needed? Yeah.Micah [00:04:40]: The story kind of begins like in 2022, 2023, like both George and I have been into AI stuff for quite a while. In 2023 specifically, I was trying to build a legal AI research assistant. So it actually worked pretty well for its era, I would say. Yeah. Yeah. So I was finding that the more you go into building something using LLMs, the more each bit of what you're doing ends up being a benchmarking problem. So had like this multistage algorithm thing, trying to figure out what the minimum viable model for each bit was, trying to optimize every bit of it as you build that out, right? Like you're trying to think about accuracy, a bunch of other metrics and performance and cost. And mostly just no one was doing anything to independently evaluate all the models. And certainly not to look at the trade-offs for speed and cost. So we basically set out just to build a thing that developers could look at to see the trade-offs between all of those things measured independently across all the models and providers. Honestly, it was probably meant to be a side project when we first started doing it.swyx [00:05:49]: Like we didn't like get together and say like, Hey, like we're going to stop working on all this stuff. I'm like, this is going to be our main thing. When I first called you, I think you hadn't decided on starting a company yet.Micah [00:05:58]: That's actually true. I don't even think we'd pause like, like George had an acquittance job. I didn't quit working on my legal AI thing. Like it was genuinely a side project.George [00:06:05]: We built it because we needed it as people building in the space and thought, Oh, other people might find it useful too. So we'll buy domain and link it to the Vercel deployment that we had and tweet about it. And, but very quickly it started getting attention. Thank you, Swyx for, I think doing an initial retweet and spotlighting it there. This project that we released. And then very quickly though, it was useful to others, but very quickly it became more useful as the number of models released accelerated. We had Mixtrel 8x7B and it was a key. That's a fun one. Yeah. Like a open source model that really changed the landscape and opened up people's eyes to other serverless inference providers and thinking about speed, thinking about cost. And so that was a key. And so it became more useful quite quickly. Yeah.swyx [00:07:02]: What I love talking to people like you who sit across the ecosystem is, well, I have theories about what people want, but you have data and that's obviously more relevant. But I want to stay on the origin story a little bit more. When you started out, I would say, I think the status quo at the time was every paper would come out and they would report their numbers versus competitor numbers. And that's basically it. And I remember I did the legwork. I think everyone has some knowledge. I think there's some version of Excel sheet or a Google sheet where you just like copy and paste the numbers from every paper and just post it up there. And then sometimes they don't line up because they're independently run. And so your numbers are going to look better than... Your reproductions of other people's numbers are going to look worse because you don't hold their models correctly or whatever the excuse is. I think then Stanford Helm, Percy Liang's project would also have some of these numbers. And I don't know if there's any other source that you can cite. The way that if I were to start artificial analysis at the same time you guys started, I would have used the Luther AI's eval framework harness. Yup.Micah [00:08:06]: Yup. That was some cool stuff. At the end of the day, running these evals, it's like if it's a simple Q&A eval, all you're doing is asking a list of questions and checking if the answers are right, which shouldn't be that crazy. But it turns out there are an enormous number of things that you've got control for. And I mean, back when we started the website. Yeah. Yeah. Like one of the reasons why we realized that we had to run the evals ourselves and couldn't just take rules from the labs was just that they would all prompt the models differently. And when you're competing over a few points, then you can pretty easily get- You can put the answer into the model. Yeah. That in the extreme. And like you get crazy cases like back when I'm Googled a Gemini 1.0 Ultra and needed a number that would say it was better than GPT-4 and like constructed, I think never published like chain of thought examples. 32 of them in every topic in MLU to run it, to get the score, like there are so many things that you- They never shipped Ultra, right? That's the one that never made it up. Not widely. Yeah. Yeah. Yeah. I mean, I'm sure it existed, but yeah. So we were pretty sure that we needed to run them ourselves and just run them in the same way across all the models. Yeah. And we were, we also did certain from the start that you couldn't look at those in isolation. You needed to look at them alongside the cost and performance stuff. Yeah.swyx [00:09:24]: Okay. A couple of technical questions. I mean, so obviously I also thought about this and I didn't do it because of cost. Yep. Did you not worry about costs? Were you funded already? Clearly not, but you know. No. Well, we definitely weren't at the start.Micah [00:09:36]: So like, I mean, we're paying for it personally at the start. There's a lot of money. Well, the numbers weren't nearly as bad a couple of years ago. So we certainly incurred some costs, but we were probably in the order of like hundreds of dollars of spend across all the benchmarking that we were doing. Yeah. So nothing. Yeah. It was like kind of fine. Yeah. Yeah. These days that's gone up an enormous amount for a bunch of reasons that we can talk about. But yeah, it wasn't that bad because you can also remember that like the number of models we were dealing with was hardly any and the complexity of the stuff that we wanted to do to evaluate them was a lot less. Like we were just asking some Q&A type questions and then one specific thing was for a lot of evals initially, we were just like sampling an answer. You know, like, what's the answer for this? Like, we didn't want to go into the answer directly without letting the models think. We weren't even doing chain of thought stuff initially. And that was the most useful way to get some results initially. Yeah.swyx [00:10:33]: And so for people who haven't done this work, literally parsing the responses is a whole thing, right? Like because sometimes the models, the models can answer any way they feel fit and sometimes they actually do have the right answer, but they just returned the wrong format and they will get a zero for that unless you work it into your parser. And that involves more work. And so, I mean, but there's an open question whether you should give it points for not following your instructions on the format.Micah [00:11:00]: It depends what you're looking at, right? Because you can, if you're trying to see whether or not it can solve a particular type of reasoning problem, and you don't want to test it on its ability to do answer formatting at the same time, then you might want to use an LLM as answer extractor approach to make sure that you get the answer out no matter how unanswered. But these days, it's mostly less of a problem. Like, if you instruct a model and give it examples of what the answers should look like, it can get the answers in your format, and then you can do, like, a simple regex.swyx [00:11:28]: Yeah, yeah. And then there's other questions around, I guess, sometimes if you have a multiple choice question, sometimes there's a bias towards the first answer, so you have to randomize the responses. All these nuances, like, once you dig into benchmarks, you're like, I don't know how anyone believes the numbers on all these things. It's so dark magic.Micah [00:11:47]: You've also got, like… You've got, like, the different degrees of variance in different benchmarks, right? Yeah. So, if you run four-question multi-choice on a modern reasoning model at the temperatures suggested by the labs for their own models, the variance that you can see on a four-question multi-choice eval is pretty enormous if you only do a single run of it and it has a small number of questions, especially. So, like, one of the things that we do is run an enormous number of all of our evals when we're developing new ones and doing upgrades to our intelligence index to bring in new things. Yeah. So, that we can dial in the right number of repeats so that we can get to the 95% confidence intervals that we're comfortable with so that when we pull that together, we can be confident in intelligence index to at least as tight as, like, a plus or minus one at a 95% confidence. Yeah.swyx [00:12:32]: And, again, that just adds a straight multiple to the cost. Oh, yeah. Yeah, yeah.George [00:12:37]: So, that's one of many reasons that cost has gone up a lot more than linearly over the last couple of years. We report a cost to run the artificial analysis. We report a cost to run the artificial analysis intelligence index on our website, and currently that's assuming one repeat in terms of how we report it because we want to reflect a bit about the weighting of the index. But our cost is actually a lot higher than what we report there because of the repeats.swyx [00:13:03]: Yeah, yeah, yeah. And probably this is true, but just checking, you don't have any special deals with the labs. They don't discount it. You just pay out of pocket or out of your sort of customer funds. Oh, there is a mix. So, the issue is that sometimes they may give you a special end point, which is… Ah, 100%.Micah [00:13:21]: Yeah, yeah, yeah. Exactly. So, we laser focus, like, on everything we do on having the best independent metrics and making sure that no one can manipulate them in any way. There are quite a lot of processes we've developed over the last couple of years to make that true for, like, the one you bring up, like, right here of the fact that if we're working with a lab, if they're giving us a private endpoint to evaluate a model, that it is totally possible. That what's sitting behind that black box is not the same as they serve on a public endpoint. We're very aware of that. We have what we call a mystery shopper policy. And so, and we're totally transparent with all the labs we work with about this, that we will register accounts not on our own domain and run both intelligence evals and performance benchmarks… Yeah, that's the job. …without them being able to identify it. And no one's ever had a problem with that. Because, like, a thing that turns out to actually be quite a good… …good factor in the industry is that they all want to believe that none of their competitors could manipulate what we're doing either.swyx [00:14:23]: That's true. I never thought about that. I've been in the database data industry prior, and there's a lot of shenanigans around benchmarking, right? So I'm just kind of going through the mental laundry list. Did I miss anything else in this category of shenanigans? Oh, potential shenanigans.Micah [00:14:36]: I mean, okay, the biggest one, like, that I'll bring up, like, is more of a conceptual one, actually, than, like, direct shenanigans. It's that the things that get measured become things that get targeted by labs that they're trying to build, right? Exactly. So that doesn't mean anything that we should really call shenanigans. Like, I'm not talking about training on test set. But if you know that you're going to be great at another particular thing, if you're a researcher, there are a whole bunch of things that you can do to try to get better at that thing that preferably are going to be helpful for a wide range of how actual users want to use the thing that you're building. But will not necessarily work. Will not necessarily do that. So, for instance, the models are exceptional now at answering competition maths problems. There is some relevance of that type of reasoning, that type of work, to, like, how we might use modern coding agents and stuff. But it's clearly not one for one. So the thing that we have to be aware of is that once an eval becomes the thing that everyone's looking at, scores can get better on it without there being a reflection of overall generalized intelligence of these models. Getting better. That has been true for the last couple of years. It'll be true for the next couple of years. There's no silver bullet to defeat that other than building new stuff to stay relevant and measure the capabilities that matter most to real users. Yeah.swyx [00:15:58]: And we'll cover some of the new stuff that you guys are building as well, which is cool. Like, you used to just run other people's evals, but now you're coming up with your own. And I think, obviously, that is a necessary path once you're at the frontier. You've exhausted all the existing evals. I think the next point in history that I have for you is AI Grant that you guys decided to join and move here. What was it like? I think you were in, like, batch two? Batch four. Batch four. Okay.Micah [00:16:26]: I mean, it was great. Nat and Daniel are obviously great. And it's a really cool group of companies that we were in AI Grant alongside. It was really great to get Nat and Daniel on board. Obviously, they've done a whole lot of great work in the space with a lot of leading companies and were extremely aligned. With the mission of what we were trying to do. Like, we're not quite typical of, like, a lot of the other AI startups that they've invested in.swyx [00:16:53]: And they were very much here for the mission of what we want to do. Did they say any advice that really affected you in some way or, like, were one of the events very impactful? That's an interesting question.Micah [00:17:03]: I mean, I remember fondly a bunch of the speakers who came and did fireside chats at AI Grant.swyx [00:17:09]: Which is also, like, a crazy list. Yeah.George [00:17:11]: Oh, totally. Yeah, yeah, yeah. There was something about, you know, speaking to Nat and Daniel about the challenges of working through a startup and just working through the questions that don't have, like, clear answers and how to work through those kind of methodically and just, like, work through the hard decisions. And they've been great mentors to us as we've built artificial analysis. Another benefit for us was that other companies in the batch and other companies in AI Grant are pushing the capabilities. Yeah. And I think that's a big part of what AI can do at this time. And so being in contact with them, making sure that artificial analysis is useful to them has been fantastic for supporting us in working out how should we build out artificial analysis to continue to being useful to those, like, you know, building on AI.swyx [00:17:59]: I think to some extent, I'm mixed opinion on that one because to some extent, your target audience is not people in AI Grants who are obviously at the frontier. Yeah. Do you disagree?Micah [00:18:09]: To some extent. To some extent. But then, so a lot of what the AI Grant companies are doing is taking capabilities coming out of the labs and trying to push the limits of what they can do across the entire stack for building great applications, which actually makes some of them pretty archetypical power users of artificial analysis. Some of the people with the strongest opinions about what we're doing well and what we're not doing well and what they want to see next from us. Yeah. Yeah. Because when you're building any kind of AI application now, chances are you're using a whole bunch of different models. You're maybe switching reasonably frequently for different models and different parts of your application to optimize what you're able to do with them at an accuracy level and to get better speed and cost characteristics. So for many of them, no, they're like not commercial customers of ours, like we don't charge for all our data on the website. Yeah. They are absolutely some of our power users.swyx [00:19:07]: So let's talk about just the evals as well. So you start out from the general like MMU and GPQA stuff. What's next? How do you sort of build up to the overall index? What was in V1 and how did you evolve it? Okay.Micah [00:19:22]: So first, just like background, like we're talking about the artificial analysis intelligence index, which is our synthesis metric that we pulled together currently from 10 different eval data sets to give what? We're pretty much the same as that. Pretty confident is the best single number to look at for how smart the models are. Obviously, it doesn't tell the whole story. That's why we published the whole website of all the charts to dive into every part of it and look at the trade-offs. But best single number. So right now, it's got a bunch of Q&A type data sets that have been very important to the industry, like a couple that you just mentioned. It's also got a couple of agentic data sets. It's got our own long context reasoning data set and some other use case focused stuff. As time goes on. The things that we're most interested in that are going to be important to the capabilities that are becoming more important for AI, what developers are caring about, are going to be first around agentic capabilities. So surprise, surprise. We're all loving our coding agents and how the model is going to perform like that and then do similar things for different types of work are really important to us. The linking to use cases to economically valuable use cases are extremely important to us. And then we've got some of the. Yeah. These things that the models still struggle with, like working really well over long contexts that are not going to go away as specific capabilities and use cases that we need to keep evaluating.swyx [00:20:46]: But I guess one thing I was driving was like the V1 versus the V2 and how bad it was over time.Micah [00:20:53]: Like how we've changed the index to where we are.swyx [00:20:55]: And I think that reflects on the change in the industry. Right. So that's a nice way to tell that story.Micah [00:21:00]: Well, V1 would be completely saturated right now. Almost every model coming out because doing things like writing the Python functions and human evil is now pretty trivial. It's easy to forget, actually, I think how much progress has been made in the last two years. Like we obviously play the game constantly of like the today's version versus last week's version and the week before and all of the small changes in the horse race between the current frontier and who has the best like smaller than 10B model like right now this week. Right. And that's very important to a lot of developers and people and especially in this particular city of San Francisco. But when you zoom out a couple of years ago, literally most of what we were doing to evaluate the models then would all be 100% solved by even pretty small models today. And that's been one of the key things, by the way, that's driven down the cost of intelligence at every tier of intelligence. We can talk about more in a bit. So V1, V2, V3, we made things harder. We covered a wider range of use cases. And we tried to get closer to things developers care about as opposed to like just the Q&A type stuff that MMLU and GPQA represented. Yeah.swyx [00:22:12]: I don't know if you have anything to add there. Or we could just go right into showing people the benchmark and like looking around and asking questions about it. Yeah.Micah [00:22:21]: Let's do it. Okay. This would be a pretty good way to chat about a few of the new things we've launched recently. Yeah.George [00:22:26]: And I think a little bit about the direction that we want to take it. And we want to push benchmarks. Currently, the intelligence index and evals focus a lot on kind of raw intelligence. But we kind of want to diversify how we think about intelligence. And we can talk about it. But kind of new evals that we've kind of built and partnered on focus on topics like hallucination. And we've got a lot of topics that I think are not covered by the current eval set that should be. And so we want to bring that forth. But before we get into that.swyx [00:23:01]: And so for listeners, just as a timestamp, right now, number one is Gemini 3 Pro High. Then followed by Cloud Opus at 70. Just 5.1 high. You don't have 5.2 yet. And Kimi K2 Thinking. Wow. Still hanging in there. So those are the top four. That will date this podcast quickly. Yeah. Yeah. I mean, I love it. I love it. No, no. 100%. Look back this time next year and go, how cute. Yep.George [00:23:25]: Totally. A quick view of that is, okay, there's a lot. I love it. I love this chart. Yeah.Micah [00:23:30]: This is such a favorite, right? Yeah. And almost every talk that George or I give at conferences and stuff, we always put this one up first to just talk about situating where we are in this moment in history. This, I think, is the visual version of what I was saying before about the zooming out and remembering how much progress there's been. If we go back to just over a year ago, before 01, before Cloud Sonnet 3.5, we didn't have reasoning models or coding agents as a thing. And the game was very, very different. If we go back even a little bit before then, we're in the era where, when you look at this chart, open AI was untouchable for well over a year. And, I mean, you would remember that time period well of there being very open questions about whether or not AI was going to be competitive, like full stop, whether or not open AI would just run away with it, whether we would have a few frontier labs and no one else would really be able to do anything other than consume their APIs. I am quite happy overall that the world that we have ended up in is one where... Multi-model. Absolutely. And strictly more competitive every quarter over the last few years. Yeah. This year has been insane. Yeah.George [00:24:42]: You can see it. This chart with everything added is hard to read currently. There's so many dots on it, but I think it reflects a little bit what we felt, like how crazy it's been.swyx [00:24:54]: Why 14 as the default? Is that a manual choice? Because you've got service now in there that are less traditional names. Yeah.George [00:25:01]: It's models that we're kind of highlighting by default in our charts, in our intelligence index. Okay.swyx [00:25:07]: You just have a manually curated list of stuff.George [00:25:10]: Yeah, that's right. But something that I actually don't think every artificial analysis user knows is that you can customize our charts and choose what models are highlighted. Yeah. And so if we take off a few names, it gets a little easier to read.swyx [00:25:25]: Yeah, yeah. A little easier to read. Totally. Yeah. But I love that you can see the all one jump. Look at that. September 2024. And the DeepSeek jump. Yeah.George [00:25:34]: Which got close to OpenAI's leadership. They were so close. I think, yeah, we remember that moment. Around this time last year, actually.Micah [00:25:44]: Yeah, yeah, yeah. I agree. Yeah, well, a couple of weeks. It was Boxing Day in New Zealand when DeepSeek v3 came out. And we'd been tracking DeepSeek and a bunch of the other global players that were less known over the second half of 2024 and had run evals on the earlier ones and stuff. I very distinctly remember Boxing Day in New Zealand, because I was with family for Christmas and stuff, running the evals and getting back result by result on DeepSeek v3. So this was the first of their v3 architecture, the 671b MOE.Micah [00:26:19]: And we were very, very impressed. That was the moment where we were sure that DeepSeek was no longer just one of many players, but had jumped up to be a thing. The world really noticed when they followed that up with the RL working on top of v3 and R1 succeeding a few weeks later. But the groundwork for that absolutely was laid with just extremely strong base model, completely open weights that we had as the best open weights model. So, yeah, that's the thing that you really see in the game. But I think that we got a lot of good feedback on Boxing Day. us on Boxing Day last year.George [00:26:48]: Boxing Day is the day after Christmas for those not familiar.George [00:26:54]: I'm from Singapore.swyx [00:26:55]: A lot of us remember Boxing Day for a different reason, for the tsunami that happened. Oh, of course. Yeah, but that was a long time ago. So yeah. So this is the rough pitch of AAQI. Is it A-A-Q-I or A-A-I-I? I-I. Okay. Good memory, though.Micah [00:27:11]: I don't know. I'm not used to it. Once upon a time, we did call it Quality Index, and we would talk about quality, performance, and price, but we changed it to intelligence.George [00:27:20]: There's been a few naming changes. We added hardware benchmarking to the site, and so benchmarks at a kind of system level. And so then we changed our throughput metric to, we now call it output speed, and thenswyx [00:27:32]: throughput makes sense at a system level, so we took that name. Take me through more charts. What should people know? Obviously, the way you look at the site is probably different than how a beginner might look at it.Micah [00:27:42]: Yeah, that's fair. There's a lot of fun stuff to dive into. Maybe so we can hit past all the, like, we have lots and lots of emails and stuff. The interesting ones to talk about today that would be great to bring up are a few of our recent things, I think, that probably not many people will be familiar with yet. So first one of those is our omniscience index. So this one is a little bit different to most of the intelligence evils that we've run. We built it specifically to look at the embedded knowledge in the models and to test hallucination by looking at when the model doesn't know the answer, so not able to get it correct, what's its probability of saying, I don't know, or giving an incorrect answer. So the metric that we use for omniscience goes from negative 100 to positive 100. Because we're simply taking off a point if you give an incorrect answer to the question. We're pretty convinced that this is an example of where it makes most sense to do that, because it's strictly more helpful to say, I don't know, instead of giving a wrong answer to factual knowledge question. And one of our goals is to shift the incentive that evils create for models and the labs creating them to get higher scores. And almost every evil across all of AI up until this point, it's been graded by simple percentage correct as the main metric, the main thing that gets hyped. And so you should take a shot at everything. There's no incentive to say, I don't know. So we did that for this one here.swyx [00:29:22]: I think there's a general field of calibration as well, like the confidence in your answer versus the rightness of the answer. Yeah, we completely agree. Yeah. Yeah.George [00:29:31]: On that. And one reason that we didn't do that is because. Or put that into this index is that we think that the, the way to do that is not to ask the models how confident they are.swyx [00:29:43]: I don't know. Maybe it might be though. You put it like a JSON field, say, say confidence and maybe it spits out something. Yeah. You know, we have done a few evils podcasts over the, over the years. And when we did one with Clementine of hugging face, who maintains the open source leaderboard, and this was one of her top requests, which is some kind of hallucination slash lack of confidence calibration thing. And so, Hey, this is one of them.Micah [00:30:05]: And I mean, like anything that we do, it's not a perfect metric or the whole story of everything that you think about as hallucination. But yeah, it's pretty useful and has some interesting results. Like one of the things that we saw in the hallucination rate is that anthropics Claude models at the, the, the very left-hand side here with the lowest hallucination rates out of the models that we've evaluated amnesty is on. That is an interesting fact. I think it probably correlates with a lot of the previously, not really measured vibes stuff that people like about some of the Claude models. Is the dataset public or what's is it, is there a held out set? There's a hell of a set for this one. So we, we have published a public test set, but we we've only published 10% of it. The reason is that for this one here specifically, it would be very, very easy to like have data contamination because it is just factual knowledge questions. We would. We'll update it at a time to also prevent that, but with yeah, kept most of it held out so that we can keep it reliable for a long time. It leads us to a bunch of really cool things, including breakdown quite granularly by topic. And so we've got some of that disclosed on the website publicly right now, and there's lots more coming in terms of our ability to break out very specific topics. Yeah.swyx [00:31:23]: I would be interested. Let's, let's dwell a little bit on this hallucination one. I noticed that Haiku hallucinates less than Sonnet hallucinates less than Opus. And yeah. Would that be the other way around in a normal capability environments? I don't know. What's, what do you make of that?George [00:31:37]: One interesting aspect is that we've found that there's not really a, not a strong correlation between intelligence and hallucination, right? That's to say that the smarter the models are in a general sense, isn't correlated with their ability to, when they don't know something, say that they don't know. It's interesting that Gemini three pro preview was a big leap over here. Gemini 2.5. Flash and, and, and 2.5 pro, but, and if I add pro quickly here.swyx [00:32:07]: I bet pro's really good. Uh, actually no, I meant, I meant, uh, the GPT pros.George [00:32:12]: Oh yeah.swyx [00:32:13]: Cause GPT pros are rumored. We don't know for a fact that it's like eight runs and then with the LM judge on top. Yeah.George [00:32:20]: So we saw a big jump in, this is accuracy. So this is just percent that they get, uh, correct and Gemini three pro knew a lot more than the other models. And so big jump in accuracy. But relatively no change between the Google Gemini models, between releases. And the hallucination rate. Exactly. And so it's likely due to just kind of different post-training recipe, between the, the Claude models. Yeah.Micah [00:32:45]: Um, there's, there's driven this. Yeah. You can, uh, you can partially blame us and how we define intelligence having until now not defined hallucination as a negative in the way that we think about intelligence.swyx [00:32:56]: And so that's what we're changing. Uh, I know many smart people who are confidently incorrect.George [00:33:02]: Uh, look, look at that. That, that, that is very humans. Very true. And there's times and a place for that. I think our view is that hallucination rate makes sense in this context where it's around knowledge, but in many cases, people want the models to hallucinate, to have a go. Often that's the case in coding or when you're trying to generate newer ideas. One eval that we added to artificial analysis is, is, is critical point and it's really hard, uh, physics problems. Okay.swyx [00:33:32]: And is it sort of like a human eval type or something different or like a frontier math type?George [00:33:37]: It's not dissimilar to frontier frontier math. So these are kind of research questions that kind of academics in the physics physics world would be able to answer, but models really struggled to answer. So the top score here is not 9%.swyx [00:33:51]: And when the people that, that created this like Minway and, and, and actually off via who was kind of behind sweep and what organization is this? Oh, is this, it's Princeton.George [00:34:01]: Kind of range of academics from, from, uh, different academic institutions, really smart people. They talked about how they turn the models up in terms of the temperature as high temperature as they can, where they're trying to explore kind of new ideas in physics as a, as a thought partner, just because they, they want the models to hallucinate. Um, yeah, sometimes it's something new. Yeah, exactly.swyx [00:34:21]: Um, so not right in every situation, but, um, I think it makes sense, you know, to test hallucination in scenarios where it makes sense. Also, the obvious question is, uh, this is one of. Many that there is there, every lab has a system card that shows some kind of hallucination number, and you've chosen to not, uh, endorse that and you've made your own. And I think that's a, that's a choice. Um, totally in some sense, the rest of artificial analysis is public benchmarks that other people can independently rerun. You provide it as a service here. You have to fight the, well, who are we to, to like do this? And your, your answer is that we have a lot of customers and, you know, but like, I guess, how do you converge the individual?Micah [00:35:08]: I mean, I think, I think for hallucinations specifically, there are a bunch of different things that you might care about reasonably, and that you'd measure quite differently, like we've called this a amnesty and solutionation rate, not trying to declare the, like, it's humanity's last hallucination. You could, uh, you could have some interesting naming conventions and all this stuff. Um, the biggest picture answer to that. It's something that I actually wanted to mention. Just as George was explaining, critical point as well is, so as we go forward, we are building evals internally. We're partnering with academia and partnering with AI companies to build great evals. We have pretty strong views on, in various ways for different parts of the AI stack, where there are things that are not being measured well, or things that developers care about that should be measured more and better. And we intend to be doing that. We're not obsessed necessarily with that. Everything we do, we have to do entirely within our own team. Critical point. As a cool example of where we were a launch partner for it, working with academia, we've got some partnerships coming up with a couple of leading companies. Those ones, obviously we have to be careful with on some of the independent stuff, but with the right disclosure, like we're completely comfortable with that. A lot of the labs have released great data sets in the past that we've used to great success independently. And so it's between all of those techniques, we're going to be releasing more stuff in the future. Cool.swyx [00:36:26]: Let's cover the last couple. And then we'll, I want to talk about your trends analysis stuff, you know? Totally.Micah [00:36:31]: So that actually, I have one like little factoid on omniscience. If you go back up to accuracy on omniscience, an interesting thing about this accuracy metric is that it tracks more closely than anything else that we measure. The total parameter count of models makes a lot of sense intuitively, right? Because this is a knowledge eval. This is the pure knowledge metric. We're not looking at the index and the hallucination rate stuff that we think is much more about how the models are trained. This is just what facts did they recall? And yeah, it tracks parameter count extremely closely. Okay.swyx [00:37:05]: What's the rumored size of GPT-3 Pro? And to be clear, not confirmed for any official source, just rumors. But rumors do fly around. Rumors. I get, I hear all sorts of numbers. I don't know what to trust.Micah [00:37:17]: So if you, if you draw the line on omniscience accuracy versus total parameters, we've got all the open ways models, you can squint and see that likely the leading frontier models right now are quite a lot bigger than the ones that we're seeing right now. And the one trillion parameters that the open weights models cap out at, and the ones that we're looking at here, there's an interesting extra data point that Elon Musk revealed recently about XAI that for three trillion parameters for GROK 3 and 4, 6 trillion for GROK 5, but that's not out yet. Take those together, have a look. You might reasonably form a view that there's a pretty good chance that Gemini 3 Pro is bigger than that, that it could be in the 5 to 10 trillion parameters. To be clear, I have absolutely no idea, but just based on this chart, like that's where you would, you would land if you have a look at it. Yeah.swyx [00:38:07]: And to some extent, I actually kind of discourage people from guessing too much because what does it really matter? Like as long as they can serve it as a sustainable cost, that's about it. Like, yeah, totally.George [00:38:17]: They've also got different incentives in play compared to like open weights models who are thinking to supporting others in self-deployment for the labs who are doing inference at scale. It's I think less about total parameters in many cases. When thinking about inference costs and more around number of active parameters. And so there's a bit of an incentive towards larger sparser models. Agreed.Micah [00:38:38]: Understood. Yeah. Great. I mean, obviously if you're a developer or company using these things, not exactly as you say, it doesn't matter. You should be looking at all the different ways that we measure intelligence. You should be looking at cost to run index number and the different ways of thinking about token efficiency and cost efficiency based on the list prices, because that's all it matters.swyx [00:38:56]: It's not as good for the content creator rumor mill where I can say. Oh, GPT-4 is this small circle. Look at GPT-5 is this big circle. And then there used to be a thing for a while. Yeah.Micah [00:39:07]: But that is like on its own, actually a very interesting one, right? That is it just purely that chances are the last couple of years haven't seen a dramatic scaling up in the total size of these models. And so there's a lot of room to go up properly in total size of the models, especially with the upcoming hardware generations. Yes.swyx [00:39:29]: So, you know. Taking off my shitposting face for a minute. Yes. Yes. At the same time, I do feel like, you know, especially coming back from Europe, people do feel like Ilya is probably right that the paradigm is doesn't have many more orders of magnitude to scale out more. And therefore we need to start exploring at least a different path. GDPVal, I think it's like only like a month or so old. I was also very positive when it first came out. I actually talked to Tejo, who was the lead researcher on that. Oh, cool. And you have your own version.George [00:39:59]: It's a fantastic. It's a fantastic data set. Yeah.swyx [00:40:01]: And maybe it will recap for people who are still out of it. It's like 44 tasks based on some kind of GDP cutoff that's like meant to represent broad white collar work that is not just coding. Yeah.Micah [00:40:12]: Each of the tasks have a whole bunch of detailed instructions, some input files for a lot of them. It's within the 44 is divided into like two hundred and twenty two to five, maybe subtasks that are the level of that we run through the agenda. And yeah, they're really interesting. I will say that it doesn't. It doesn't necessarily capture like all the stuff that people do at work. No avail is perfect is always going to be more things to look at, largely because in order to make the tasks well enough to find that you can run them, they need to only have a handful of input files and very specific instructions for that task. And so I think the easiest way to think about them are that they're like quite hard take home exam tasks that you might do in an interview process.swyx [00:40:56]: Yeah, for listeners, it is not no longer like a long prompt. It is like, well, here's a zip file with like a spreadsheet or a PowerPoint deck or a PDF and go nuts and answer this question.George [00:41:06]: OpenAI released a great data set and they released a good paper which looks at performance across the different web chat bots on the data set. It's a great paper, encourage people to read it. What we've done is taken that data set and turned it into an eval that can be run on any model. So we created a reference agentic harness that can run. Run the models on the data set, and then we developed evaluator approach to compare outputs. That's kind of AI enabled, so it uses Gemini 3 Pro Preview to compare results, which we tested pretty comprehensively to ensure that it's aligned to human preferences. One data point there is that even as an evaluator, Gemini 3 Pro, interestingly, doesn't do actually that well. So that's kind of a good example of what we've done in GDPVal AA.swyx [00:42:01]: Yeah, the thing that you have to watch out for with LLM judge is self-preference that models usually prefer their own output, and in this case, it was not. Totally.Micah [00:42:08]: I think the way that we're thinking about the places where it makes sense to use an LLM as judge approach now, like quite different to some of the early LLM as judge stuff a couple of years ago, because some of that and MTV was a great project that was a good example of some of this a while ago was about judging conversations and like a lot of style type stuff. Here, we've got the task that the grader and grading model is doing is quite different to the task of taking the test. When you're taking the test, you've got all of the agentic tools you're working with, the code interpreter and web search, the file system to go through many, many turns to try to create the documents. Then on the other side, when we're grading it, we're running it through a pipeline to extract visual and text versions of the files and be able to provide that to Gemini, and we're providing the criteria for the task and getting it to pick which one more effectively meets the criteria of the task. Yeah. So we've got the task out of two potential outcomes. It turns out that we proved that it's just very, very good at getting that right, matched with human preference a lot of the time, because I think it's got the raw intelligence, but it's combined with the correct representation of the outputs, the fact that the outputs were created with an agentic task that is quite different to the way the grading model works, and we're comparing it against criteria, not just kind of zero shot trying to ask the model to pick which one is better.swyx [00:43:26]: Got it. Why is this an ELO? And not a percentage, like GDP-VAL?George [00:43:31]: So the outputs look like documents, and there's video outputs or audio outputs from some of the tasks. It has to make a video? Yeah, for some of the tasks. Some of the tasks.swyx [00:43:43]: What task is that?George [00:43:45]: I mean, it's in the data set. Like be a YouTuber? It's a marketing video.Micah [00:43:49]: Oh, wow. What? Like model has to go find clips on the internet and try to put it together. The models are not that good at doing that one, for now, to be clear. It's pretty hard to do that with a code editor. I mean, the computer stuff doesn't work quite well enough and so on and so on, but yeah.George [00:44:02]: And so there's no kind of ground truth, necessarily, to compare against, to work out percentage correct. It's hard to come up with correct or incorrect there. And so it's on a relative basis. And so we use an ELO approach to compare outputs from each of the models between the task.swyx [00:44:23]: You know what you should do? You should pay a contractor, a human, to do the same task. And then give it an ELO and then so you have, you have human there. It's just, I think what's helpful about GDPVal, the OpenAI one, is that 50% is meant to be normal human and maybe Domain Expert is higher than that, but 50% was the bar for like, well, if you've crossed 50, you are superhuman. Yeah.Micah [00:44:47]: So we like, haven't grounded this score in that exactly. I agree that it can be helpful, but we wanted to generalize this to a very large number. It's one of the reasons that presenting it as ELO is quite helpful and allows us to add models and it'll stay relevant for quite a long time. I also think it, it can be tricky looking at these exact tasks compared to the human performance, because the way that you would go about it as a human is quite different to how the models would go about it. Yeah.swyx [00:45:15]: I also liked that you included Lama 4 Maverick in there. Is that like just one last, like...Micah [00:45:20]: Well, no, no, no, no, no, no, it is the, it is the best model released by Meta. And... So it makes it into the homepage default set, still for now.George [00:45:31]: Other inclusion that's quite interesting is we also ran it across the latest versions of the web chatbots. And so we have...swyx [00:45:39]: Oh, that's right.George [00:45:40]: Oh, sorry.swyx [00:45:41]: I, yeah, I completely missed that. Okay.George [00:45:43]: No, not at all. So that, which has a checkered pattern. So that is their harness, not yours, is what you're saying. Exactly. And what's really interesting is that if you compare, for instance, Claude 4.5 Opus using the Claude web chatbot, it performs worse than the model in our agentic harness. And so in every case, the model performs better in our agentic harness than its web chatbot counterpart, the harness that they created.swyx [00:46:13]: Oh, my backwards explanation for that would be that, well, it's meant for consumer use cases and here you're pushing it for something.Micah [00:46:19]: The constraints are different and the amount of freedom that you can give the model is different. Also, you like have a cost goal. We let the models work as long as they want, basically. Yeah. Do you copy paste manually into the chatbot? Yeah. Yeah. That's, that was how we got the chatbot reference. We're not going to be keeping those updated at like quite the same scale as hundreds of models.swyx [00:46:38]: Well, so I don't know, talk to a browser base. They'll, they'll automate it for you. You know, like I have thought about like, well, we should turn these chatbot versions into an API because they are legitimately different agents in themselves. Yes. Right. Yeah.Micah [00:46:53]: And that's grown a huge amount of the last year, right? Like the tools. The tools that are available have actually diverged in my opinion, a fair bit across the major chatbot apps and the amount of data sources that you can connect them to have gone up a lot, meaning that your experience and the way you're using the model is more different than ever.swyx [00:47:10]: What tools and what data connections come to mind when you say what's interesting, what's notable work that people have done?Micah [00:47:15]: Oh, okay. So my favorite example on this is that until very recently, I would argue that it was basically impossible to get an LLM to draft an email for me in any useful way. Because most times that you're sending an email, you're not just writing something for the sake of writing it. Chances are context required is a whole bunch of historical emails. Maybe it's notes that you've made, maybe it's meeting notes, maybe it's, um, pulling something from your, um, any of like wherever you at work store stuff. So for me, like Google drive, one drive, um, in our super base databases, if we need to do some analysis or some data or something, preferably model can be plugged into all of those things and can go do some useful work based on it. The things that like I find most impressive currently that I am somewhat surprised work really well in late 2025, uh, that I can have models use super base MCP to query read only, of course, run a whole bunch of SQL queries to do pretty significant data analysis. And. And make charts and stuff and can read my Gmail and my notion. And okay. You actually use that. That's good. That's, that's, that's good. Is that a cloud thing? To various degrees of order, but chat GPD and Claude right now, I would say that this stuff like barely works in fairness right now. Like.George [00:48:33]: Because people are actually going to try this after they hear it. If you get an email from Micah, odds are it wasn't written by a chatbot.Micah [00:48:38]: So, yeah, I think it is true that I have never actually sent anyone an email drafted by a chatbot. Yet.swyx [00:48:46]: Um, and so you can, you can feel it right. And yeah, this time, this time next year, we'll come back and see where it's going. Totally. Um, super base shout out another famous Kiwi. Uh, I don't know if you've, you've any conversations with him about anything in particular on AI building and AI infra.George [00:49:03]: We have had, uh, Twitter DMS, um, with, with him because we're quite big, uh, super base users and power users. And we probably do some things more manually than we should in. In, in super base support line because you're, you're a little bit being super friendly. One extra, um, point regarding, um, GDP Val AA is that on the basis of the overperformance of the models compared to the chatbots turns out, we realized that, oh, like our reference harness that we built actually white works quite well on like gen generalist agentic tasks. This proves it in a sense. And so the agent harness is very. Minimalist. I think it follows some of the ideas that are in Claude code and we, all that we give it is context management capabilities, a web search, web browsing, uh, tool, uh, code execution, uh, environment. Anything else?Micah [00:50:02]: I mean, we can equip it with more tools, but like by default, yeah, that's it. We, we, we give it for GDP, a tool to, uh, view an image specifically, um, because the models, you know, can just use a terminal to pull stuff in text form into context. But to pull visual stuff into context, we had to give them a custom tool, but yeah, exactly. Um, you, you can explain an expert. No.George [00:50:21]: So it's, it, we turned out that we created a good generalist agentic harness. And so we, um, released that on, on GitHub yesterday. It's called stirrup. So if people want to check it out and, and it's a great, um, you know, base for, you know, generalist, uh, building a generalist agent for more specific tasks.Micah [00:50:39]: I'd say the best way to use it is get clone and then have your favorite coding. Agent make changes to it, to do whatever you want, because it's not that many lines of code and the coding agents can work with it. Super well.swyx [00:50:51]: Well, that's nice for the community to explore and share and hack on it. I think maybe in, in, in other similar environments, the terminal bench guys have done, uh, sort of the Harbor. Uh, and so it's, it's a, it's a bundle of, well, we need our minimal harness, which for them is terminus and we also need the RL environments or Docker deployment thing to, to run independently. So I don't know if you've looked at it. I don't know if you've looked at the harbor at all, is that, is that like a, a standard that people want to adopt?George [00:51:19]: Yeah, we've looked at it from a evals perspective and we love terminal bench and, and host benchmarks of, of, of terminal mention on artificial analysis. Um, we've looked at it from a, from a coding agent perspective, but could see it being a great, um, basis for any kind of agents. I think where we're getting to is that these models have gotten smart enough. They've gotten better, better tools that they can perform better when just given a minimalist. Set of tools and, and let them run, let the model control the, the agentic workflow rather than using another framework that's a bit more built out that tries to dictate the, dictate the flow. Awesome.swyx [00:51:56]: Let's cover the openness index and then let's go into the report stuff. Uh, so that's the, that's the last of the proprietary art numbers, I guess. I don't know how you sort of classify all these. Yeah.Micah [00:52:07]: Or call it, call it, let's call it the last of like the, the three new things that we're talking about from like the last few weeks. Um, cause I mean, there's a, we do a mix of stuff that. Where we're using open source, where we open source and what we do and, um, proprietary stuff that we don't always open source, like long context reasoning data set last year, we did open source. Um, and then all of the work on performance benchmarks across the site, some of them, we looking to open source, but some of them, like we're constantly iterating on and so on and so on and so on. So there's a huge mix, I would say, just of like stuff that is open source and not across the side. So that's a LCR for people. Yeah, yeah, yeah, yeah.swyx [00:52:41]: Uh, but let's, let's, let's talk about open.Micah [00:52:42]: Let's talk about openness index. This. Here is call it like a new way to think about how open models are. We, for a long time, have tracked where the models are open weights and what the licenses on them are. And that's like pretty useful. That tells you what you're allowed to do with the weights of a model, but there is this whole other dimension to how open models are. That is pretty important that we haven't tracked until now. And that's how much is disclosed about how it was made. So transparency about data, pre-training data and post-training data. And whether you're allowed to use that data and transparency about methodology and training code. So basically, those are the components. We bring them together to score an openness index for models so that you can in one place get this full picture of how open models are.swyx [00:53:32]: I feel like I've seen a couple other people try to do this, but they're not maintained. I do think this does matter. I don't know what the numbers mean apart from is there a max number? Is this out of 20?George [00:53:44]: It's out of 18 currently, and so we've got an openness index page, but essentially these are points, you get points for being more open across these different categories and the maximum you can achieve is 18. So AI2 with their extremely open OMO3 32B think model is the leader in a sense.swyx [00:54:04]: It's hooking face.George [00:54:05]: Oh, with their smaller model. It's coming soon. I think we need to run, we need to get the intelligence benchmarks right to get it on the site.swyx [00:54:12]: You can't have it open in the next. We can not include hooking face. We love hooking face. We'll have that, we'll have that up very soon. I mean, you know, the refined web and all that stuff. It's, it's amazing. Or is it called fine web? Fine web. Fine web.Micah [00:54:23]: Yeah, yeah, no, totally. Yep. One of the reasons this is cool, right, is that if you're trying to understand the holistic picture of the models and what you can do with all the stuff the company's contributing, this gives you that picture. And so we are going to keep it up to date alongside all the models that we do intelligence index on, on the site. And it's just an extra view to understand.swyx [00:54:43]: Can you scroll down to this? The, the, the, the trade-offs chart. Yeah, yeah. That one. Yeah. This, this really matters, right? Obviously, because you can b

The top AI news from the past week, every ThursdAI

Ho Ho Ho, Alex here! (a real human writing these words, this needs to be said in 2025) Merry Christmas (to those who celebrate) and welcome to the very special yearly ThursdAI recap! This was an intense year in the world of AI, and after 51 weekly episodes (this is episode 52!) we have the ultimate record of all the major and most important AI releases of this year! So instead of bringing you a weekly update (it's been a slow week so far, most AI labs are taking a well deserved break, the Cchinese AI labs haven't yet surprised anyone), I'm dropping a comprehensive yearly AI review! Quarter by quarter, month by month, both in written form and as a pod/video! Why do this? Who even needs this? Isn't most of it obsolete? I have asked myself this exact question while prepping for the show (it was quite a lot of prep, even with Opus's help). I eventually landed on, hey, if nothing else, this will serve as a record of the insane week of AI progress we all witnessed. Can you imagine that the term Vibe Coding is less than 1 year old? That Claude Code was released at the start of THIS year? We get hedonicly adapt to new AI goodies so quick, and I figured this will serve as a point in time check, we can get back to and feel the acceleration! With that, let's dive in - P.S. the content below is mostly authored by my co-author for this, Opus 4.5 high, which at the end of 2025 I find the best creative writer with the best long context coherence that can imitate my voice and tone (hey, I'm also on a break!

Dame Rueda
247. Especial: EICMA Parte 2 (Naked - Sporturismo)

Dame Rueda

Play Episode Listen Later Dec 22, 2025 119:57


Seguimos desgranando todo lo importante del EICMA 2025 en esta segunda parte del especial, centrada en las novedades que no son trail, porque de eso ya hubo ración suficiente . Hablamos a fondo de: Naked deportivas y neoretro, con especial atención a las propuestas más interesantes y a las que generan polémica. Honda y su ofensiva: CB1000F, CB1000 GT y el espectacular prototipo V3 con compresor eléctrico, uno de los grandes focos del salón. CFMoto, que vuelve a demostrar que ya no es solo “la china barata”, con conceptos muy serios y cada vez más ambiciosos. Suzuki, y el eterno debate sobre los refritos, los precios y hasta dónde estiran una plataforma. Comparativas reales con modelos rivales, análisis de posicionamiento, precios conocidos (y los que no), y opinión sin filtros. Además: Debate técnico (del bueno) sobre electrónica, ayudas a la conducción y sensaciones. Reflexión sobre hacia dónde va la moto actual y qué se está perdiendo (o ganando) por el camino. Feedback de oyentes, con opiniones enfrentadas, acuerdos, palos y risas. Anuncios y novedades del equipo de Dame Rueda de cara a próximos eventos.

KagoshimaniaX
姶良市がV3達成、指宿市も急上昇!データで見る鹿児島の「住みここち」最新事情

KagoshimaniaX

Play Episode Listen Later Dec 17, 2025 0:26


「姶良市がV3達成、指宿市も急上昇!データで見る鹿児島の「住みここち」最新事情」 こんにちは!カゴシマニアックス編集部です。 皆さんは今住んでいる街に「愛着」や「誇り」を持っていますか? 「買い物に便利だから」「自然が豊かだから」「昔から住んでいるから」など、住む場所を選ぶ理由は人それぞれですよね。 […]

Starforged: Tabula Rasa
Harrison Potterson & The Shiny Rock V1.1 - Part 1 of 2

Starforged: Tabula Rasa

Play Episode Listen Later Dec 15, 2025 57:43


An edit of Wizard People, Dear Reader by Brad Neely with the care of broadcast network TV on rated R movies being played on daytime weekends in the 90s.For the V3.0.1 with an excellent soundtrack, see⁠Https://jadedharmacabal.bandcamp.com/album/harrison-potterson-the-shiny-rock-ver-301⁠On YouTube here:⁠ https://www.youtube.com/watch?v=8MXkFcxLqtM

tv rock shiny v3 dear reader brad neely wizard people
Starforged: Tabula Rasa
Harrison Potterson & The Shiny Rock V1.1.1 - Part 2 of 2

Starforged: Tabula Rasa

Play Episode Listen Later Dec 15, 2025 54:48


An edit of Wizard People, Dear Reader by Brad Neely with the care of broadcast network TV on rated R movies being played on daytime weekends in the 90s.For the V3.0.1 with an excellent soundtrack, see⁠Https://jadedharmacabal.bandcamp.com/album/harrison-potterson-the-shiny-rock-ver-301⁠Or on YouTube here: https://www.youtube.com/watch?v=h1t5CvLdk_g

tv rock shiny v3 dear reader brad neely wizard people
Normal Nerds
One Piece Episode 1151 Breaks the Internet — Bonney Awakens! MHA Season 8 Coming to End, Gachiakuta's Best Fight Yet, OPM Improves

Normal Nerds

Play Episode Listen Later Dec 4, 2025 50:52


This week on the Normal Nerds Podcast, Maxwell and Davis deliver a loaded Weekly Anime Recap.They kick things off with My Hero Academia Season 8, Episode 9, talking Deku's questionable haircut, Bakugo refusing to rest, and the Todoroki family's heavy, emotional wrap-up—plus why Endeavor's redemption hits harder than the fandom gives him credit for.Next, the boys rave about Spy x Family Season 3, Episode 9 with Double Starlight Anya, Damien sprinting to her rescue, Becky's chaos, Melinda's unhinged emotional whiplash, and Anya's ego skyrocketing into villain-cape territory.Then they tear into One Punch Man Season 3, Episode 8—a rare episode with decent animation that STILL can't save the season. Flashy Flash's fight pops off, Kid Emperor is actually interesting, but the stiff faces and still frames keep dragging the whole show down.From there, it's pure hype with Tojima Wants to Be a Kamen Rider Episode 9. Tojima lands the punch heard 'round the world, Rider Man goes feral, V3 gives respect where it's due, and the show continues to be one of the season's biggest surprises.They follow it up with the comfy, consistently fun May I Ask You One Last Thing? Episode 10, featuring Scarlet punching monsters into paste, Julius shutting down betrayals instantly, dragon-guy simps going wild, and a war brewing in the background.Then they officially consider dropping SANDA Episode 9, after another episode that leans way too hard into the bizarre. With the series' tone spiraling and its creative lineage showing, they debate if it's even worth continuing.After the main block, they jump into two big highlights of the week:Gachiakuta Episode 21 delivers insane action as Bundus takes on multiple Cleaners at once, Rudo evolves his 3R into a railgun-level weapon, Zodyl reveals his real plan, and Noerde's terrifying return sets up massive consequences.One Piece Episode 1151 goes OFF: Gear 5 Luffy returns, Bonney awakens a Nika-style form of her own, and the Iron Giant finally steps in—setting the stage for one of Egghead's most explosive episodes yet.If you're into big analysis, big fights, and big laughs, this is one of the most packed recaps the guys have ever done.Support the show

✨Poki - Podcast over Kunstmatige Intelligentie AI
Alexander kan eindelijk programmeren + DeepSeek haalt Amerikaanse modellen in + de 'ziel' van Claude gelekt

✨Poki - Podcast over Kunstmatige Intelligentie AI

Play Episode Listen Later Dec 4, 2025 70:59


Alexander heeft een doorbraak: als niet-programmeur bouwde hij in één dag een werkende tool met Claude Opus 4.5. Cursor heeft nu een agent mode die alle code verbergt - je ziet alleen een chatbalk en een browservenster. Ondertussen koopt Anthropic Bun over, cruciaal stuk infrastructuur. De puzzelstukjes vallen op hun plek: straks bouw je custom tools op het web en neemt Anthropic een percentage van je omzet.DeepSeek is terug met V3.2, bijna even goed als GPT-5 en Gemini 3. Wat we niet wisten: je gebruikt Chinese modellen waarschijnlijk al zonder het te weten - bedrijven prototypen met gesloten modellen en vervangen die daarna door open source. En ook lanceert Mistral tien nieuwe Europese modellen als alternatief.En toen lekte daar ook nog het ‘soul document' van Claude: 14.000 woorden over hoe hun AI moet denken, zich moet gedragen én voelen. Iedere zin zit vol nuance - wees lief maar niet te lief, je lijkt op een mens maar bent geen mens. Dit alles doet denken: wanneer komt die AGI nou? Dwarkesh en Sutskever denken 10 tot 20 jaar. Het bewijs: Anthropic traint nu specifiek op Excel in plaats van generaliseerbare intelligentie.Als je een lezing wil over AI van Wietse of Alexander dan kan dat. Mail ons op lezing@aireport.emailVandaag nog beginnen met AI binnen jouw bedrijf? Ga dan naar deptagency.com/aireport This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.aireport.email/subscribe

Die Hupe | Auto- und Motorrad-Nerdcast
Folge 68: Kia EV6 GT, Tesla Model 3 und Motorradhighlights auf der EICMA 2025

Die Hupe | Auto- und Motorrad-Nerdcast

Play Episode Listen Later Dec 1, 2025 66:47 Transcription Available


Wir sprechen über unsere Messe-Highlights der EICMA 2025: Vom elektrischen V3 bei Honda bis hin zu Chinas neuem Motorrad-Feuerwerk und vielversprechenden E-Großrollern. Außerdem: Clemens im Kia EV6 GT, Sebastian hat sich in seine Lotus Elise neu verliebt und das (vorläufige) Ende der Tesla-Odyssee.

The Instagram Stories
11-26-25 - Does Reposting Your Own Content on Instagram Do Anything?

The Instagram Stories

Play Episode Listen Later Nov 27, 2025 11:38


YouTube tests a new custom feed just like Instagram and TikTok, Shorts AI Creation Tools get updates, and I experiment with speeding up Lauren to see if that helps. Also the Head of Instagram stops by to explain that reposting your own content to Feed really won't do much, and the team at TikTok shares some stats around using Creators to make content for brands. After the music, I do Wednesday Waffle talking about a book I read recently. Links:YouTube: Testing "Your Custom Feed" (Google Support: YouTube)YouTube: New Communities Features, Expansion of Shorts AI Creation Tools, and Handles Update! (YouTube)Instagram: Does Reposting Your Content To Feed Help? (Instagram)TikTok: The Creator Advantage: How creators drive real brand impact on TikTok (TikTok)TikTok: TikTok One - Creative Academy Videos (TikTok) Wednesday Waffle:Book: Wrong Place, Wrong Time - Gillian McAllister (Amazon) Transcript: Daniel Hill: Welcome to the Instagram stories for Wednesday, November 26th. I'm your host, Daniel Hill. There is a lot of social media news to talk about today. The YouTube team has expanded their shorts AI creation tools. The head of Instagram explains whether or not it's worth it to repost your own posts and if that'll help you get more engagement. The team at TikTok shares some data explaining why it's so important to work with creators if you're a business and how that can drive brand impact. for your business. We'll get into that along with some video guides that the TikTok team has made to help you make better content. And after all the social media news, I will do a Wednesday waffle where I talk about a topic that may or may not be related to social media. All of that and more on today's episode. But first, here's a quick word from our sponsors. Welcome back. Let's start with the YouTube news. Before we dive in, a little bit of context. I've been talking on the show recently about how Instagram is going to allow you to customize what you see in your for you feed based on what you are personally interested in and you can pick which topics you're interested in, which ones you're not. Tik Tok has had that for a long time. There are sliders that you can move to indicate which kinds of content you want to see more or less of. And they recently added the ability to adjust what level of AI content you see in your feed. Now, YouTube is copying that and they shared yesterday that they are testing something called your custom feed. They say, quote, "We're experimenting presenting with a new feature called your custom feed that lets you customize recommendations for your home feed. If you are part of the experiment, you will see your custom feed appear on your homepage as a chip beside home. When you click into it, you can update your existing home feed recommendations by entering a single prompt. This feature is designed to give you an easy to use way to have more control over your suggested content. If you see it, check it out and share your feedback". I will link to this post in the show notes so that you can see it for yourself. All right, moving on. Since we're already talking about YouTube news, let's move to Lauren from the YouTube Creator Insider team with her updates talking about how the Shorts AI creation tools are being expanded and an update to the way YouTube handles channel names versus handles. Here's the clip from Lauren. Uh, one quick thing before I play the clip. I was reading the feedback that I got about the show and some of you mentioned that Lauren's updates can drag on a little bit and I agree. So, I'm going to experiment with speeding up Lauren just a little bit. Hopefully, it's enough that it goes faster and you don't feel like Lauren's dragging, but you can still catch what she's saying.Lauren: What's up, insiders? I'm Lauren, a program manager working on our product team here at YouTube and the producer of Creator Insider. Up until now, channel names were used as the identifier for channels across live chat and channel memberships on main and YouTube studio. Now, a creator's handle will be shown across these services as their identifier. For moderators of live chat, you can still navigate to a user's channel by tapping on their handle. Let us know if you have any questions. In June, we talked about new AI powered shorts creation tools. If you missed the update, we'll leave more information in the description. We're happy to share that we're expanding standalone clips, green screen backgrounds, AI playground, and phototovideo to new markets around the world for creators with their YouTube language settings set to English. We're also leveling up the photo to video experience with new prompt capabilities. Now you can create a prompt from scratch, watch your memories come to life, and even add speech to give your video a voice. We're also introducing new Genai effects that transform your sketches into captivating videos powered by VO. These effects are now available globally. Additionally, speech to song and the ability to add lyrics and vocals in Dream Track are now available to creators in the US. These features will be rolling out this month and we'll keep you posted as we add new features. We're also bringing the power of Google DeepMind's V3 model to shorts, available for everyone on mobile. This upgrade from V2 lets you create videos up to 8 seconds long, previously six, now with synchronized sound effects, ambient audio, and speech. We'll leave more info below. Next, updates for communities. If you're still on the fence about enabling communities, an internal experiment in early September 2025 found that channels with YouTube communities enabled saw on average an increase in post impressions and likes on their channel.Daniel Hill: Okay, I'm going to stop it there because the rest of the update is about communities and I don't think it's very interesting. But if you do want to check out the whole post, I will link to it in the show notes so you can watch it for yourself. Hopefully the increased speed with which Lauren explained those things still let you understand what was going on and hopefully kept her a little bit more brief than usual. Okay, now let's move on to the Instagram section, the head of Instagram answered the question about whether or not reposting your content in your own feed does anything. So, you have the opportunity to share content that you've made from your feed to your story, for instance, but now you can also repost it to your feed. If you posted a piece of content and it didn't really do that well, it might be tempting to repost it to your feed so that your followers have another chance to see it. The head of Instagram explains it's not really worth it to do that. Here's the clip.Adam Mosseri: Since we launched reposts a couple months ago. I get the question a lot. Should I repost my own content? And you can. It might help a little bit on the margins, but it's not going to meaningfully change the amount of reach that you get. If you want to try and help your post go a little bit farther, I'd recommend instead going into the comments, responding to some people, liking some comments, and interacting with the people who've taken the time to actually like or comment on your post. This will help more than just reposting something that you've already posted. But I understand why people try. And this not going to hurt you to do so, but it's not going to actually help. So, I wanted to answer that question definitively once and for all. Hopefully, this helps later.Daniel Hill: So, there you have it. Not really worth a lot of time and energy. We're going to take a quick break. When we come back, some information from the Tik Tok team about how creators can help to drive impact for brands and additionally some videos from the Tik Tok team helping you to make better content. Stick around. Welcome back. Let's continue with the Tik Tok news. The team at Tik Tok made a long blog post sharing some data about how much creators making content and having brands push that content can impact the business that the brand does as opposed to the brand just making content on their own or hiring a marketing company. The importance of this cannot be understated because the content comes across as more authentic. They share some stats explaining that creator ads meaning an ad that is based on a piece of content that a creator made that creator ad can drive a 70% higher click-through rate and 159% higher engagement rate than noncreator ads. Okay, so why such a big difference? Three main reasons. First, when creators are making content, they're doing it through the lens of Tik Tok culture. They're familiar with the platform, not from the perspective of trying to sell a product or service, but rather just being familiar with the community. Additionally, creators can make a lot of good content very quickly. We are all used to sitting down to come up with an idea of something that we think could potentially work, coming up with what we need in order to make that piece of content, whether it's a script, finding a location, then filming it, editing it, and publishing it, and doing that for ourselves. So, when brands are working with creators, they're tapping into this system that we are all doing all the time. Anyway, another key thing to remember is that when brands partner with creators, the Those creators have a wide variety of different voices, skill sets, things they bring to the table, all of which appeal to different people. So, it does make sense to partner with a wide variety of creators. The third reason that this is so effective is because people already follow these creators. They liked them enough to follow them. When a creator makes a piece of content about a brand or product or service and posts it to their account, it comes across more authentically because it is not coming from the advertiser's account. According to a study from the TikTok team, ads posted to a creator's account have a 59% higher engagement rate and a 16% higher 6second viewthrough rate than those that are not posted directly to the creator's account. So, it's worth it to do this. There's more stats and strategy in this blog post, which I will link to, but I also uncovered something called TikTok 1, which is a creative academy to help you make better content on TikTok. I will play a short snippet of one of the videos that I found on there so you can get an idea of what this is all about.Unknown Speaker (from TikTok 1 Clip): This video is about TikTok creative best practices. Good creative is imperative for successful ad campaign. There are some essential guidelines you must follow in order to set yourself up for that success. These include video duration, design elements, safe zone, and video formats. Today we're talking creative codes. Six secrets to help you decode TikTok's creative potential. And what better place to start then code number one Tik Tok first. So when we say TikTok first, what do we mean? Going TikTok first means creating natural feeling TikTok content that's authentic to the platform. Feeling authentic to the for you page is as simple as taking cues from the content you love. How can you make content that feels organic to the for you page? Here are some quick tips that will help you look right at home on TikTok. Start simple. From filming at a professional shoot to filming on your phone, you can execute your ideas in the way that works for you. Go 9x6. This is a platform where vertical video thrives. Frame your content accordingly. Shoot high-res. Whether you use high-end software or smartphone technology, create video content that is clear and crisp.Daniel Hill: Okay, I'm going to stop the clip there. You get the idea. If you want to watch the rest of this video series, which I actually think is very good, I will link to it in the show notes. Be sure to check it out for yourself. That is it for today's news. If you would like to hear me do Wednesday waffle, which is where I talk about another topic that may or may not be related to social media. Stick around after the music.Music: Instagram news got you covered. Sometimes even TikTok relevant platforms in the metaverse. Ahead of the wave without a break or a pit stops waiting for Zuckerberg to give me the big job. Use trademarks and logos with instance permission. Of course, if you like the show and you gain some good info, maybe leave a review. It's a type of applause. Just drop me a message if you want to collab. If you got some good content or you want to run ads at @DanielHillMedia is where I am. TikTok, Facebook, at Instagram. All right, thank you for sticking around to hear me talk about something else. And now I would like to recommend a book. I don't think I did this previously on the show. I would like to recommend a book called Wrong Place, Wrong Time by Gillian McAllister. I would categorize this book as a thriller/time travel book. I don't like to normally recommend books because everyone likes different things in books. However, this book is excellent. I spend a lot of time reading books. I like to read a few minutes every night before I go to sleep, and this one really had me hooked. I was having trouble forcing myself actually to go to sleep. The ending ties together perfectly. I will link to this book in the show notes. Definitely check it out. Wrong place, Wrong Time by Gillian McAllister. If you like time travel, if you like science fiction, if you like thrillers, if you like any books that are a bit mysterious or suspenseful, I think you will like this. Find the link to the book in the show notes. And thank you for listening to me talk about something else other than social media for a minute.Sign Up for The Weekly Email Roundup: NewsletterLeave a Review: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠Apple Podcasts⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠Follow Me on Instagram: ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠@danielhillmedia⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠ Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.

Midjourney : Fast Hours
Go Flux Yourself: Midjourney Still Runs the Realism Game

Midjourney : Fast Hours

Play Episode Listen Later Nov 16, 2025 59:05


Drew and Rory start with eyeball horror, Stranger Things hype, and the idea of AI-powered contact lenses before stumbling straight into the real mind-melt: Midjourney, Grok Imagine, Mystic 3, and Flux all colliding in one episode. They roast their own prompts, trigger an accidental NSF-DoubleU moment live inside Grok, argue about “flux face,” and still somehow manage to pull out real, practical tips for people trying to make better AI images without losing their minds.Across an hour of chaos, they unpack Midjourney v8's subtle shifts, hidden personalization signals, Style Explorer tricks, Smart Search shortcuts, Grok's Sora-style infinite feed, Mystic 3's scary-good skin detail, and why Midjourney still owns lo-fi, lived-in, “shot-on-a-phone” energy. If you care about composition, cinematic ratios, editorial portraits, food realism, or just want to hear two people dunk on Flux and node editors while actually teaching you something, this one hits.Listeners will come away knowing how to use stills archive for composition, when to skip upscales for more analog realism, how Grok Imagine's image + video workflow really behaves, and where Mystic 3 can replace Midjourney in a serious portrait or product stack.--⏱️ Midjourney Fast Hour0:00 Intro, eyeballs, and a Friday brain check2:05 Contact lens horror stories, Mission Impossible, Black Mirror eyes3:07 Stranger Things Season 5 hype and binge vs weekly TV4:51 Movies, biopics, sports docs, and couch season setting in6:23 Cowboys documentary, sports pipelines, and TV as passive story feed7:00 AI overload, nobody keeping up, and why this pod exists8:30 Midjourney profiles, Style Creator, and new personalization talk9:29 Like/dislike buttons as hidden training data and 7:3 aspect ratio love10:35 Stills Archive, cinematic framing, and cleaner compositions12:00 Style Explorer vs old-school SREF and what quietly vanished13:16 Three under-the-radar Midjourney Smart Search + right-click + Option-upscale tweaks15:35 V8, fewer wall-of-text prompts, and a move toward visual controls18:12 First look at Grok Imagine's interface and infinite scroll feel19:35 Sora-style endless bottom feed, variants, and “make video” in Grok22:51 Cinematic looks, color grading, and Grok as “idea and curate” engine24:19 Live NSFW surprise inside Grok Imagine and instant rating change25:23 Finding Grok history, stills, and video exports with sound26:31 Who actually gets Grok video and Drew's first real reaction to using it27:38 Mystic 3 enters the chat and upscaling less for analog vibes29:02 Why “too sharp” screams AI and how grain + smart detail saves realism30:18 Outpainting, editing, and why Midjourney still wins surgical compositing35:01 Mystic 3 V3 screen-share and first impressions35:45 Editorial portraits, skin detail, eyelashes, and hands that finally look human37:26 Mystic 3 model comparisons: Zen, State-of-the-Art, and weird description blur39:16 Zooming all the way into pores, fingerprints, and micro skin texture43:44 Cocktail and food prompts where Mystic falls behind Midjourney50:05 Nano Banana 2 rumors, native 4K wishes, and how Midjourney might respond50:58 Why Midjourney still rules lo-fi, disposable camera, and Polaroid-style shots52:16 Grok Imagine vs Flux vs Midjourney for lived-in Y2K flash photos53:39 Flux face, direct flash tests, and “go flux yourself” is born55:30 Nodes, Grok workflows, and why scrolling is faster than wiring graphs56:01 Why Midjourney is avoiding node-based interfaces on purpose57:05 Final sendoff: go flux yourself and get out of here

LAB: The Podcast
LAB the Podcast: Poetry Corner with Wendy Kieffer | The Patience of Ordinary Things

LAB: The Podcast

Play Episode Listen Later Nov 14, 2025 39:00


In this episode of LAB the Podcast, poet and V3 artist Wendy Kieffer joins us for a conversation inspired by “The Patience of Ordinary Things” by Pat Schneider. From cups that hold tea to floors that receive our feet, Wendy and Zach reflect on the quiet love hidden in the simplest moments — and how seeing with fresh eyes can awaken gratitude, imagination, and joy.Take a pause from the hurry, pour a warm cup of tea, and join us as we reflect on patience, presence, and the extraordinary grace of ordinary thingsThank you for joining the conversation and embodying the life and beauty of the gospel. Don't forget to like, subscribe, and follow LAB the Podcast. Support / Sponsor: https://vuvivo.com/supportFor More Videos, Subscribe: @VUVIVOV3 | https://www.youtube.com/@VUVIVOV3Follow: @labthepodcast | @vuvivo_v3 | @zachjelliott | @wendy.kiefferLike:  https://www.facebook.com/vuvivo.v3Order Alchemy of Praise: https://www.amazon.com/dp/1944470220?ref=cm_sw_r_cp_ud_dp_3NCER8Y41NXPRQ5QE469&ref_=cm_sw_r_cp_ud_dp_3NCER8Y41NXPRQ5QE469&social_share=cm_sw_r_cp_ud_dp_3NCER8Y41NXPRQ5QE469&skipTwisterOG=2Support the show

朝日新聞 ニュースの現場から
Wリーグ、外国籍選手の枠拡大 男子日本代表候補を分析(バスクラ)#2062

朝日新聞 ニュースの現場から

Play Episode Listen Later Nov 14, 2025 51:42


【SoftBank ウインターカップ2025 観戦チケットキャンペーン中】https://ciy.digital.asahi.com/ciy/11017934 (12/14締め切り) 【番組内容】Wリーグは、外国籍選手の枠が拡大しました。これまでは「通算5年以上日本に在留」が条件でしたが、今シーズンから撤廃になりました。また、「FIBAバスケットボールワールドカップ2027アジア地区予選 window1」の男子日本代表の候補メンバーについても語ります。 ※2025年11月5日に収録しました。バスケ通信―クラッチタイム(バスクラ)の過去回はこちら( https://bit.ly/49fLo67 )。 プレイリスト( https://buff.ly/4iNnkOj ) 【2カ月間無料キャンペーン】全記事が読み放題!デジタル版30周年を記念し、11/25まで実施中★後日、有料会員限定の報談スペシャルバージョンを配信します!https://digital.asahi.com/pr/cp/2025/aut/?ref=cp2025aut_podcast_gaiyoran 【関連記事】富士通、V3へ白星発進 バスケ・女子Wリーグ開幕https://www.asahi.com/articles/DA3S16326605.html?iref=omny (有料会員の方はログインしていただくと、過去の紙面記事検索として閲覧できます)千葉Jの19歳、瀬川琉久が狙うエースの座「勇樹さんから先発奪う」https://www.asahi.com/articles/ASTB235ZYTB2UTQP025M.html?iref=omny 10季目迎えたBリーグ、地方でも熱気 隣県対決の平日チケット完売https://www.asahi.com/articles/ASTBQ0RRBTBQTIPE009M.html?iref=omny   【出演・スタッフ】松本龍三郎(スポーツ部) https://x.com/asahi_bballinfo / https://buff.ly/lpil4Q8 松本麻美(メディア事業本部・スポーツ事業部) https://x.com/Asa_asa_sports MC・音源編集 堀江麻友 https://bit.ly/4kepWoO 【おねがい】朝日新聞ポッドキャストは、みなさまからの購読料で配信しています。番組継続のため、会員登録をお願いします! http://t.asahi.com/womz 【朝ポキ情報】アプリで記者と対話 http://t.asahi.com/won1 交流はdiscord https://bit.ly/asapoki_discord おたよりフォーム https://bit.ly/asapoki_otayori 朝ポキTV https://www.youtube.com/@asapoki_official メルマガ https://bit.ly/asapoki_newsletter 広告ご検討の企業様は http://t.asahi.com/asapokiguide 番組検索ツール https://bit.ly/asapoki_cast 最新情報はX https://bit.ly/asapoki_twitter See omnystudio.com/listener for privacy information.

Ethereum Daily - Crypto News Briefing
Balancer V2 Suffers $110m Exploit

Ethereum Daily - Crypto News Briefing

Play Episode Listen Later Nov 4, 2025 4:16


Balancer V2 suffers a $110m exploit. The EF ESP team reopens grant applications. StarkWare launches the S-two prover on Starknet. And ZKP2P releases V3 of its onramp protocol. Read more: https://ethdaily.io/815 Disclaimer: Content is for informational purposes only, not endorsement or investment advice. The accuracy of information is not guaranteed.

LAB: The Podcast
LAB the Podcast with Cammie Elliott: Trusting God in All Things

LAB: The Podcast

Play Episode Listen Later Oct 24, 2025 49:12


In this special episode of LAB the Podcast, we sit down with Zach's favorite guest—and favorite person—Cammie Elliott, Director of Ministry for VU VI VO.Cammie is the heartbeat behind so much of V3's work—from LAB the Podcast to the Immersive Experiences, Sehnsucht Symphony and all special projects. Together, Zach and Cammie explore what it means to trust God in every season, from cross-country moves and moments of loss to the daily rhythms of faith, prayer, and building a home that reflects the life and beauty of the Gospel.They share about learning to hear and obey, cultivating beauty in the ordinary, and discovering God's character through His names. If you've ever struggled to trust God with what you can't see—this conversation will inspire you to rest in who He is.Thank you for joining the conversation and embodying the life and beauty of the gospel. Don't forget to like, subscribe, and follow LAB the Podcast. Support / SponsorLearn more about V3For More Videos, Subscribe: @VUVIVOV3 | YouTubeFollow: @labthepodcast | @vuvivo_v3 | @zachjelliott Support the show

LAB: The Podcast
LAB the Podcast with Liza Thurmond: Seeing the Gospel with Fresh Eyes

LAB: The Podcast

Play Episode Listen Later Oct 17, 2025 59:05


Liza Thurmond joins LAB the Podcast to talk about the sacred work of teaching, seeing beauty through faith, and how one trip to Chicago's Art Institute reshaped her classroom and church. A powerful conversation about V3's ON BEAUTY Immersive Experience at the Art Institute of Chicago and how it gave her fresh eyes. Thank you for joining the conversation and embodying the life and beauty of the gospel. Don't forget to like, subscribe, and follow LAB the Podcast. Register for On Beauty: REGISTERSupport / SponsorFor More Videos, Subscribe: @VUVIVOV3 | YouTubeFollow: @labthepodcast | @vuvivo_v3 | @zachjelliott Support the show

Mechanista in G – Scanline Media
Mechanista in G – Anchor

Mechanista in G – Scanline Media

Play Episode Listen Later Oct 15, 2025


Sometimes even Gundam Crossbone has good ideas. And they may not be good ideas leading up to them, or even following them. But for one shining moment, you get a cartridge-powered pile bunker leg driving a massive beam saber down, and everything is ok. ...Ok, that's unfair to the F89, that thing's fine. But boy. Anchor V2, V3, V4... it definitely goes downhill. Lighting can strike twice. But it usually doesn't. dont fact check dont fact check dont fact check You can find a video version of this podcast for free on Scanline Media's Patreon! If you want to find us on Bluesky, Dylan is lowpolyrobot.bsky.social and Six is six.scanlinemedia.com. Our opening theme is the Hangar Theme from Gundam Breaker 3, and our ending theme for this episode is Resumption from Gundam Breaker 4. Our podcast art is a fantastic piece of work from Twitter artist @fenfelt. Want to see a list of every unit we've covered from every episode, including variants and tangents? It's right here. Units discussed: F89 Gundam F89 Cristonian Ferley Danganloid MegaZord Gamma Kajimerian Ben G Doggie Kruger (Dekaranger) Doggie Kruger (SPD) F89 Gundam F89 (Final Equipment Type) F89 Gundam F89 (Chaser Equipment Type) F89 Gundam F89 (Commander Specifications) Anchor Anchor V2 Anchor V3 Anchor V4

The Defiant
The $10 Trillion Question: Crypto's Biggest Cycle Yet with Ellio Trades

The Defiant

Play Episode Listen Later Aug 25, 2025 45:53


In this episode of The Defiant Podcast, we sit down with Ellio Trades, Co-Founder of BlackHole, the fastest-growing decentralized exchange on Avalanche. Ellio shares his insights on the current state of the crypto market, including whether we're entering a new altcoin season or witnessing a slower, more sustainable growth cycle. He dives deep into the maturation of the crypto space, the role of institutional investors, and why this cycle could be the most significant in crypto history.Ellio also unpacks the innovative mechanics behind BlackHole, from its V3.3 DEX model to its focus on creating deep liquidity and sustainable revenue. He explains how BlackHole is reshaping the DeFi landscape and why Avalanche was the perfect ecosystem for its launch. Watch now to gain valuable insights on the future of blockchain, decentralized finance, and the next wave of crypto adoption.Chapters00:00 The $10 Trillion Question: Is Altcoin Season Here?00:04 Traditional Cycles vs. A New Crypto Paradigm00:42 The Maturation of DeFi and Institutional Adoption01:30 Introducing Ellio Trades and BlackHole02:33 Bitcoin, Ethereum, and the Shift to Regulated Markets03:45 Why Ethereum's Value Proposition is Finally Clicking05:18 The Role of Wall Street in Crypto's Next Big Cycle07:32 Balancing Supercycle Narratives with Traditional Cycles10:33 DeFi's Undervalued Potential and Institutional Interest12:23 Stablecoins and the Future of Blockchain Applications16:04 Why BlackHole Chose Avalanche for Its Launch18:42 BlackHole's Explosive Growth and Unique Features23:03 How BlackHole's V3.3 Model Creates Deep Liquidity30:28 Lessons from Past Projects and BlackHole's Innovations35:02 The Long-Term Vision for BlackHole and DeFi

AI Unraveled: Latest AI News & Trends, Master GPT, Gemini, Generative AI, LLMs, Prompting, GPT Store

A daily Chronicle of AI Innovations August 22nd 2025:Listen at https://podcasts.apple.com/us/podcast/ai-daily-rundown-aug-22-2025-google-analyzes-geminis/id1684415169?i=1000723151588Hello AI Unraveled Listeners,In today's AI News,

Kitesurf365 | a podcast for kitesurfers
EXCLUSIVE Megaloop List Revealed + Slingshot Machine V3 | The Megapod

Kitesurf365 | a podcast for kitesurfers

Play Episode Listen Later Aug 21, 2025 45:13


  On today's episode, we discuss the last six men selected from the video entries, and we hear from the six ladies who will compete in their first Red Bull Megaloop. We call Yücel Paralik and congratulate him on selection. We also hear from Jeremy Burlando regarding the Slingshot Machine V3, and Colin and Adrian talk about their WOO teams.   Slingshot machine V3   https://slingshotsports.com/collections/kites/products/machine-v3    Portrait:   https://portraitkite.com     https://www.fantasykite.com   Woo Sports:   https://woosports.com   Follow us:   https://www.instagram.com/portraitkite/   https://www.instagram.com/kitesurf365/

Henry Lake
What is "The Playbook" and why have they partnered with V3 Sports?

Henry Lake

Play Episode Listen Later Aug 19, 2025 12:03


Henry talks with the Founder and CEO of The Playbook, Dr. T M Robinson-Mosley about the reason for the app, their partnership with national sports leagues, why they wanted to be a part of the V3, their "21 Day Check In Challenge,"and more.

Ethereum Daily - Crypto News Briefing
Tornado Cash Trial Ends With Mixed Verdict

Ethereum Daily - Crypto News Briefing

Play Episode Listen Later Aug 6, 2025 4:03


The US vs Storm trial ends with a mixed verdict. Aave releases the V3 developer toolkit. Pendle launches a funding rate trading protocol. And Cosmos Health announces its $300 million ETH reserve strategy. Read more: https://ethdaily.io/756 Disclaimer: Content is for informational purposes only, not endorsement or investment advice. The accuracy of information is not guaranteed.

feeder sound
premiere: Cher - 34 [Art House]

feeder sound

Play Episode Listen Later Jul 28, 2025 13:08


Under his moniker Cher, Romanian artist Traian Cherecheș presents C1M3R4_Analog cut_V3 on his freshly founded label Art House, an inspiring collection of six tracks that explore the essence of minimal music. Read more @ feeder.ro/2025/07/28/cher-c1m3r4-analog-cut-v3

Bird Camp
Field Armor, a new and improved dog vest, and camp talk with Jeff

Bird Camp

Play Episode Listen Later Jul 16, 2025 59:52


Jeff wanted to get the word out about the new V3 dog vest and catch up on this falls plans. Of course there was plenty of BirdCamp shop talk as well.The GoFundMe link mentioned is here. https://www.gofundme.com/f/7zmz6-support-rachels-battle-against-breast-cancer/cl/s?attribution_id=sl:71e0677b-3f2b-4b3f-884e-c3adb3d17873&lang=en_US&utm_campaign=fp_sharesheet&utm_content=amp13_t1-amp14_t2-amp15_t1&utm_medium=customer&utm_source=copy_link&v=amp14_t2Thank you to our sponsorsAspen Thicket Grouse Dogs aspenthicketgrousedogs.comPine Hill Gun Dogs phkscllc@gmail.comSecond Chance Bird dogs Field Armor fieldarmorusa.comWild Card Outfitters and Guide Service wildcardoutdoors.comPrairie ridge Farms prairieridgefarms.com

The Late Night Vision Show
Ep. 375 - AGM Rattler V3 TS50-640 **EXCLUSIVE REVIEW**

The Late Night Vision Show

Play Episode Listen Later Jul 10, 2025 39:08


In this episode, Jason and Hans dive into their review of the AGM Rattler V3 50-640 LRF. The V3 is the first compact Rattler model equipped with a built in LRF. And not only does it have 1,000  meter LRF, it's built into the lens so it is tucked away and adds no bulk or weight to the scope! The V3 also includes a ballistic calculator, a 640×512, sub-15 mK thermal sensor and huge 2560×2560 OLED display.  Tune in to find out if this scope is worth the price tag for serious night hunters and learn how it compares in value to other similar scopes. 

Faith Bible Church Menifee Sermon Podcast

1 Corinthians 11:27–34 (ESV) — 27 Whoever, therefore, eats the bread or drinks the cup of the Lord in an unworthy manner will be guilty concerning the body and blood of the Lord. 28 Let a person examine himself, then, and so eat of the bread and drink of the cup. 29 For anyone who eats and drinks without discerning the body eats and drinks judgment on himself. 30 That is why many of you are weak and ill, and some have died. 31 But if we judged ourselves truly, we would not be judged. 32 But when we are judged by the Lord, we are disciplined so thatwe may not be condemned along with the world. 33 So then, my brothers, when you come together to eat, wait for one another— 34 if anyone is hungry, let him eat at home—so that when you come together it will not be for judgment. About the other things I will give directions when I come. IN COMMUNION YOU ARE TO EXAMINE, REMEMBER,PROCLAIM AND ANTICIPATE THE GOSPEL OF CHRIST!  THE NECESSITY OF PERSONAL EXAMINATION v27-30 a)    TheUnworthy Manner-      “Now, if we would catch the meaning of this declaration, we must know what it is to eat unworthily. Some restrict it to the Corinthians, and the abuse that had crept in among them, but I am of opinion that Paul here, according to his usual manner, passed on from the particular case to a general statement, or from one instance to an entire class. There was one fault that prevailed among the Corinthians. He takes occasion from this to speak of every kind of faulty administration or reception of the Supper. “God,” says he, “will not allow this sacrament to be profaned without punishing it severely.” John Calvin Commentaries on the Epistles of Paul the Apostle to the Corinthians, vol. 1 pg 385.b)    The Worthy Manner -        (v24)Thanks, (v24-25) Remembrance, (v26) Proclamation and Anticipation (Ephesians 4:1-3)  c)     The General Principles-       (v27-29) Personal Examination (whoever… let a person… himself… then eat and drink…) -        (v27, 29) Guilt that leads to judgement or participation without meditation. (YERPA)d)    The Specific Judgement -        (v30)  Some are weak, ill and died.   PERSONAL EXAMINATION UNDER THE PATERNAL LOVE OF THE FATHER v31-32  General Principles:  a)    Your Freedom: Instruction In The Gospel (v31)a.    A clean conscience (Hebrews 10:22) b.     Full confession (1 John 1:7-9) c.      A true humility (1 timothy 1:12-17) d.     Informed progress (Ephesians 4:1-3)      b)    His Faithfulness: Discipline In The Gospel (v32)Hebrews 12:3-14a.     V3-4 Considering Christ b.     V5-10 Remember the love of the Fatherc.      V11 Discipline brings perishing pain,  and progressive paternal peace and perfection… d.     V12 -14 Therefore – take action… (12) up in hope, (13) forward in healing, (14) outward in holiness   PRACTICAL CONCLUSION OF EXAMINATION   v33-34 Specific commands To Corinth and practical applicationfor us: a)     (v33) Hopeful and Patient to serve others  b)    (v34) Humble and Prepared to serve others -        More specifics in patience

LAB: The Podcast
LAB the Podcast with Riley Cooper: “Behind the Scenes”

LAB: The Podcast

Play Episode Listen Later Jun 20, 2025 62:44


He's usually the one making others shine but today we flip the script. On this episode of LAB the Podcast, we sit down with Tampa native and Podcast Producer, Riley Cooper. Riley opens up about his hometown roots, his evolving walk with God, and the winding path that brought him to V3. We also talk about the deep impact the Wayfarer Podcast had on his life and how it became a turning point in his story. You won't want to miss this behind-the-scenes look at one of our own.Thank you for joining the conversation and embodying the life and beauty of the gospel. Don't forget to like, subscribe, and follow LAB the Podcast. Support / Sponsor@VUVIVOV3 | YouTube@labthepodcast | @vuvivo_v3 | @zachjelliott | @wayfarerpodcast Support the show

AI For Humans
OpenAI Prepares For Artificial Super Intelligence, Apple's Major AI Fail & New Insane 1X Robots

AI For Humans

Play Episode Listen Later Jun 12, 2025 45:51


OpenAI's Sam Altman drops o3-Pro & sees “The Gentle Singularity”, Ilya Sutskever prepares for super intelligence & Mark Zuckerberg is spending MEGA bucks on AI talent. WHAT GIVES? All of the major AI companies are not only preparing for AGI but for true “super intelligence” which is on the way, at least according to *them*. What does that mean for us? And how do we exactly prepare for it? Also, Apple's WWDC is a big AI letdown, Eleven Labs' new V3 model is AMAZING, Midjourney got sued and, oh yeah, those weird 1X Robotics androids are back and running through grassy fields. WHAT WILL HAPPEN WHEN AI IS SMARTER THAN US? ACTUALLY, IT PROB ALREADY IS. #ai #ainews #openai Join the discord: https://discord.gg/muD2TYgC8f Join our Patreon: https://www.patreon.com/AIForHumansShow AI For Humans Newsletter: https://aiforhumans.beehiiv.com/ Follow us for more on X @AIForHumansShow Join our TikTok @aiforhumansshow To book us for speaking, please visit our website: https://www.aiforhumans.show/   // Show Links /?   Ilya Sutsketver's Commencement Speech About AI https://youtu.be/zuZ2zaotrJs?si=U_vHVpFEyTRMWSNa Apple's Cringe Genmoji Video https://x.com/altryne/status/1932127782232076560 OpenAI's Sam Altman On Superintelligence “The Gentle Singularity” https://blog.samaltman.com/the-gentle-singularity The Secret Mathematicians Meeting Where The Tried To Outsmart AI https://www.scientificamerican.com/article/inside-the-secret-meeting-where-mathematicians-struggled-to-outsmart-ai/ O3-Pro Released  https://x.com/sama/status/1932532561080975797 The most expensive o3-Pro Hello https://x.com/Yuchenj_UW/status/1932544842405720540 Eleven Labs v3  https://x.com/elevenlabsio/status/1930689774278570003 o3 regular drops in price by 80% - cheaper than GPT-4o  https://x.com/edwinarbus/status/1932534578469654552 Open weights model taking a ‘little bit more time' https://x.com/sama/status/1932573231199707168 Meta Buys 49% of Scale AI + Alexandr Wang Comes In-House https://www.nytimes.com/2025/06/10/technology/meta-new-ai-lab-superintelligence.html Apple Underwhelms at WWDC Re AI https://www.cnbc.com/2025/06/09/apple-wwdc-underwhelms-on-ai-software-biggest-facelift-in-decade-.html BusinessWeek's Mark Gurman on WWDC https://x.com/markgurman/status/1932145561919991843 Joanna Stern Grills Apple https://youtu.be/NTLk53h7u_k?si=AvnxM9wefXl2Nyjn Midjourney Sued by Disney & Comcast https://www.reuters.com/business/media-telecom/disney-universal-sue-image-creator-midjourney-copyright-infringement-2025-06-11/ 1x Robotic's Redwood https://x.com/1x_tech/status/1932474830840082498 https://www.1x.tech/discover/redwood-ai Redwood Mobility Video https://youtu.be/Dp6sqx9BGZs?si=UC09VxSx-PK77q-- Amazon Testing Humanoid Robots To Deliver Packages https://www.theinformation.com/articles/amazon-prepares-test-humanoid-robots-delivering-packages?rc=c3oojq&shared=736391f5cd5d0123 Autonomous Drone Beats Pilots For the First Time https://x.com/AISafetyMemes/status/1932465150151270644 Random GPT-4o Image Gen Pic https://www.reddit.com/r/ChatGPT/comments/1l7nnnz/what_do_you_get/?share_id=yWRAFxq3IMm9qBYxf-ZqR&utm_content=4&utm_medium=ios_app&utm_name=ioscss&utm_source=share&utm_term=1 https://x.com/AIForHumansShow/status/1932441561843093513 Jon Finger's Shoes to Cars With Luma's Modify Video https://x.com/mrjonfinger/status/1932529584442069392

In-Ear Insights from Trust Insights
In-Ear Insights: How Generative AI Reasoning Models Work

In-Ear Insights from Trust Insights

Play Episode Listen Later Jun 11, 2025


In this episode of In-Ear Insights, the Trust Insights podcast, Katie and Chris discuss the Apple AI paper and critical lessons for effective prompting, plus a deep dive into reasoning models. You’ll learn what reasoning models are and why they sometimes struggle with complex tasks, especially when dealing with contradictory information. You’ll discover crucial insights about AI’s “stateless” nature, which means every prompt starts fresh and can lead to models getting confused. You’ll gain practical strategies for effective prompting, like starting new chats for different tasks and removing irrelevant information to improve AI output. You’ll understand why treating AI like a focused, smart intern will help you get the best results from your generative AI tools. Tune in to learn how to master your AI interactions! Watch the video here: Can’t see anything? Watch it on YouTube here. Listen to the audio here: https://traffic.libsyn.com/inearinsights/tipodcast-how-generative-ai-reasoning-models-work.mp3 Download the MP3 audio here. Need help with your company’s data and analytics? Let us know! Join our free Slack group for marketers interested in analytics! [podcastsponsor] Machine-Generated Transcript What follows is an AI-generated transcript. The transcript may contain errors and is not a substitute for listening to the episode. Christopher S. Penn – 00:00 In this week’s In Ear Insights, there is so much in the AI world to talk about. One of the things that came out recently that I think is worth discussing, because we can talk about the basics of good prompting as part of it, Katie, is a paper from Apple. Apple’s AI efforts themselves have stalled a bit, showing that reasoning models, when given very complex puzzles—logic-based puzzles or spatial-based puzzles, like moving blocks from stack to stack and getting them in the correct order—hit a wall after a while and then just collapse and can’t do anything. So, the interpretation of the paper is that there are limits to what reasoning models can do and that they can kind of confuse themselves. On LinkedIn and social media and stuff, Christopher S. Penn – 00:52 Of course, people have taken this to the illogical extreme, saying artificial intelligence is stupid, nobody should use it, or artificial general intelligence will never happen. None of that is within the paper. Apple was looking at a very specific, narrow band of reasoning, called deductive reasoning. So what I thought we’d talk about today is the paper itself to a degree—not a ton about it—and then what lessons we can learn from it that will make our own AI practices better. So to start off, when we talk about reasoning, Katie, particularly you as our human expert, what does reasoning mean to the human? Katie Robbert – 01:35 When I think, if you say, “Can you give me a reasonable answer?” or “What is your reason?” Thinking about the different ways that the word is casually thrown around for humans. The way that I think about it is, if you’re looking for a reasonable answer to something, then that means that you are putting the expectation on me that I have done some kind of due diligence and I have gathered some kind of data to then say, “This is the response that I’m going to give you, and here are the justifications as to why.” So I have some sort of a data-backed thinking in terms of why I’ve given you that information. When I think about a reasoning model, Katie Robbert – 02:24 Now, I am not the AI expert on the team, so this is just my, I’ll call it, amateurish understanding of these things. So, a reasoning model, I would imagine, is similar in that you give it a task and it’s, “Okay, I’m going to go ahead and see what I have in my bank of information for this task that you’re asking me about, and then I’m going to do my best to complete the task.” When I hear that there are limitations to reasoning models, I guess my first question for you, Chris, is if these are logic problems—complete this puzzle or unfurl this ball of yarn, kind of a thing, a complex thing that takes some focus. Katie Robbert – 03:13 It’s not that AI can’t do this; computers can do those things. So, I guess what I’m trying to ask is, why can’t these reasoning models do it if computers in general can do those things? Christopher S. Penn – 03:32 So you hit on a really important point. The tasks that are in this reasoning evaluation are deterministic tasks. There’s a right and wrong answer, and what they’re supposed to test is a model’s ability to think through. Can it get to that? So a reasoning model—I think this is a really great opportunity to discuss this. And for those who are listening, this will be available on our YouTube channel. A reasoning model is different from a regular model in that it thinks things through in sort of a first draft. So I’m showing DeepSeq. There’s a button here called DeepThink, which switches models from V3, which is a non-reasoning model, to a reasoning model. So watch what happens. I’m going to type in a very simple question: “Which came first, the chicken or the egg?” Katie Robbert – 04:22 And I like how you think that’s a simple question, but that’s been sort of the perplexing question for as long as humans have existed. Christopher S. Penn – 04:32 And what you see here is this little thinking box. This thinking box is the model attempting to solve the question first in a rough draft. And then, if I had closed up, it would say, “Here is the answer.” So, a reasoning model is essentially—we call it, I call it, a hidden first-draft model—where it tries to do a first draft, evaluates its own first draft, and then produces an answer. That’s really all it is. I mean, yes, there’s some mathematics going on behind the scenes that are probably not of use to folks listening to or watching the podcast. But at its core, this is what a reasoning model does. Christopher S. Penn – 05:11 Now, if I were to take the exact same prompt, start a new chat here, and instead of turning off the deep think, what you will see is that thinking box will no longer appear. It will just try to solve it as is. In OpenAI’s ecosystem—the ChatGPT ecosystem—when you pull down that drop-down of the 82 different models that you have a choice from, there are ones that are called non-reasoning models: GPT4O, GPT4.1. And then there are the reasoning models: 0304 mini, 04 mini high, etc. OpenAI has done a great job of making it as difficult as possible to understand which model you should use. But that’s reasoning versus non-reasoning. Google, very interestingly, has moved all of their models to reasoning. Christopher S. Penn – 05:58 So, no matter what version of Gemini you’re using, it is a reasoning model because Google’s opinion is that it creates a better response. So, Apple was specifically testing reasoning models because in most tests—if I go to one of my favorite websites, ArtificialAnalysis.ai, which sort of does a nice roundup of smart models—you’ll notice that reasoning models are here. And if you want to check this out and you’re listening, ArtificialAnalysis.ai is a great benchmark set that wraps up all the other benchmarks together. You can see that the leaderboards for all the major thinking tests are all reasoning models, because that ability for a model to talk things out by itself—really having a conversation with self—leads to much better results. This applies even for something as simple as a blog post, like, “Hey, let’s write a blog post about B2B marketing.” Christopher S. Penn – 06:49 Using a reasoning model will let the model basically do its own first draft, critique itself, and then produce a better result. So that’s what a reasoning model is, and why they’re so important. Katie Robbert – 07:02 But that didn’t really answer my question, though. I mean, I guess maybe it did. And I think this is where someone like me, who isn’t as technically inclined or isn’t in the weeds with this, is struggling to understand. So I understand what you’re saying in terms of what a reasoning model is. A reasoning model, for all intents and purposes, is basically a model that’s going to talk through its responses. I’ve seen this happen in Google Gemini. When I use it, it’s, “Okay, let me see. You’re asking me to do this. Let me see what I have in the memory banks. Do I have enough information? Let me go ahead and give it a shot to answer the question.” That’s basically the synopsis of what you’re going to get in a reasoning model. Katie Robbert – 07:48 But if computers—forget AI for a second—if calculations in general can solve those logic problems that are yes or no, very black and white, deterministic, as you’re saying, why wouldn’t a reasoning model be able to solve a puzzle that only has one answer? Christopher S. Penn – 08:09 For the same reason they can’t do math, because the type of puzzle they’re doing is a spatial reasoning puzzle which requires—it does have a right answer—but generative AI can’t actually think. It is a probabilistic model that predicts based on patterns it’s seen. It’s a pattern-matching model. It’s the world’s most complex next-word prediction machine. And just like mathematics, predicting, working out a spatial reasoning puzzle is not a word problem. You can’t talk it out. You have to be able to visualize in your head, map it—moving things from stack to stack—and then coming up with the right answers. Humans can do this because we have many different kinds of reasoning: spatial reasoning, musical reasoning, speech reasoning, writing reasoning, deductive and inductive and abductive reasoning. Christopher S. Penn – 09:03 And this particular test was testing two of those kinds of reasoning, one of which models can’t do because it’s saying, “Okay, I want a blender to fry my steak.” No matter how hard you try, that blender is never going to pan-fry a steak like a cast iron pan will. The model simply can’t do it. In the same way, it can’t do math. It tries to predict patterns based on what’s been trained on. But if you’ve come up with a novel test that the model has never seen before and is not in its training data, it cannot—it literally cannot—repeat that task because it is outside the domain of language, which is what it’s predicting on. Christopher S. Penn – 09:42 So it’s a deterministic task, but it’s a deterministic task outside of what the model can actually do and has never seen before. Katie Robbert – 09:50 So then, if I am following correctly—which, I’ll be honest, this is a hard one for me to follow the thread of thinking on—if Apple published a paper that large language models can’t do this theoretically, I mean, perhaps my assumption is incorrect. I would think that the minds at Apple would be smarter than collectively, Chris, you and I, and would know this information—that was the wrong task to match with a reasoning model. Therefore, let’s not publish a paper about it. That’s like saying, “I’m going to publish a headline saying that Katie can’t run a five-minute mile; therefore, she’s going to die tomorrow, she’s out of shape.” No, I can’t run a five-minute mile. That’s a fact. I’m not a runner. I’m not physically built for it. Katie Robbert – 10:45 But now you’re publishing some kind of information about it that’s completely fake and getting people in the running industry all kinds of hyped up about it. It’s irresponsible reporting. So, I guess that’s sort of my other question. If the big minds at Apple, who understand AI better than I ever hope to, know that this is the wrong task paired with the wrong model, why are they getting us all worked up about this thing by publishing a paper on it that sounds like it’s totally incorrect? Christopher S. Penn – 11:21 There are some very cynical hot takes on this, mainly that Apple’s own AI implementation was botched so badly that they look like a bunch of losers. We’ll leave that speculation to the speculators on LinkedIn. Fundamentally, if you read the paper—particularly the abstract—one of the things they were trying to test is, “Is it true?” They did not have proof that models couldn’t do this. Even though, yes, if you know language models, you would know this task is not well suited to it in the same way that they’re really not suited to geography. Ask them what the five nearest cities to Boston are, show them a map. They cannot figure that out in the same way that you and I use actual spatial reasoning. Christopher S. Penn – 12:03 They’re going to use other forms of essentially tokenization and prediction to try and get there. But it’s not the same and it won’t give the same answers that you or I will. It’s one of those areas where, yeah, these models are very sophisticated and have a ton of capabilities that you and I don’t have. But this particular test was on something that they can’t do. That’s asking them to do complex math. They cannot do it because it’s not within the capabilities. Katie Robbert – 12:31 But I guess that’s what I don’t understand. If Apple’s reputation aside, if the data scientists at that company knew—they already knew going in—it seems like a big fat waste of time because you already know the answer. You can position it, however, it’s scientific, it’s a hypothesis. We wanted to prove it wasn’t true. Okay, we know it’s not true. Why publish a paper on it and get people all riled up? If it is a PR play to try to save face, to be, “Well, it’s not our implementation that’s bad, it’s AI in general that’s poorly constructed.” Because I would imagine—again, this is a very naive perspective on it. Katie Robbert – 13:15 I don’t know if Apple was trying to create their own or if they were building on top of an existing model and their implementation and integration didn’t work. Therefore, now they’re trying to crap all over all of the other model makers. It seems like a big fat waste of time. When I—if I was the one who was looking at the budget—I’m, “Why do we publish that paper?” We already knew the answer. That was a waste of time and resources. What are we doing? I’m genuinely, again, maybe naive. I’m genuinely confused by this whole thing as to why it exists in the first place. Christopher S. Penn – 13:53 And we don’t have answers. No one from Apple has given us any. However, what I think is useful here for those of us who are working with AI every day is some of the lessons that we can learn from the paper. Number one: the paper, by the way, did not explain particularly well why it thinks models collapsed. It actually did, I think, a very poor job of that. If you’ve worked with generative AI models—particularly local models, which are models that you run on your computer—you might have a better idea of what happened, that these models just collapsed on these reasoning tasks. And it all comes down to one fundamental thing, which is: every time you have an interaction with an AI model, these models are called stateless. They remember nothing. They remember absolutely nothing. Christopher S. Penn – 14:44 So every time you prompt a model, it’s starting over from scratch. I’ll give you an example. We’ll start here. We’ll say, “What’s the best way to cook a steak?” Very simple question. And it’s going to spit out a bunch of text behind the scenes. And I’m showing my screen here for those who are listening. You can see the actual prompt appearing in the text, and then it is generating lots of answers. I’m going to stop that there just for a moment. And now I’m going to ask the same question: “Which came first, the chicken or the egg?” Christopher S. Penn – 15:34 The history of the steak question is also part of the prompt. So, I’ve changed conversation. You and I, in a chat or a text—group text, whatever—we would just look at the most recent interactions. AI doesn’t do that. It takes into account everything that is in the conversation. So, the reason why these models collapsed on these tasks is because they were trying to solve it. And when they’re thinking aloud, remember that first draft we showed? All of the first draft language becomes part of the next prompt. So if I said to you, Katie, “Let me give you some directions on how to get to my house.” First, you’re gonna take a right, then you take a left, and then you’re gonna go straight for two miles, and take a right, and then. Christopher S. Penn – 16:12 Oh, wait, no—actually, no, there’s a gas station. Left. No, take a left there. No, take a right there, and then go another two miles. If I give you those instructions, which are full of all these back twists and turns and contradictions, you’re, “Dude, I’m not coming over.” Katie Robbert – 16:26 Yeah, I’m not leaving my house for that. Christopher S. Penn – 16:29 Exactly. Katie Robbert – 16:29 Absolutely not. Christopher S. Penn – 16:31 Absolutely. And that’s what happens when these reasoning models try to reason things out. They fill up their chat with so many contradicting answers as they try to solve the problem that on the next turn, guess what? They have to reprocess everything they’ve talked about. And so they just get lost. Because they’re reading the whole conversation every time as though it was a new conversation. They’re, “I don’t know what’s going on.” You said, “Go left,” but they said, “Go right.” And so they get lost. So here’s the key thing to remember when you’re working with any generative AI tool: you want to keep as much relevant stuff in the conversation as possible and remove or eliminate irrelevant stuff. Christopher S. Penn – 17:16 So it’s a really bad idea, for example, to have a chat where you’re saying, “Let’s write a blog post about B2B marketing.” And then say, “Oh, I need to come up with an ideal customer profile.” Because all the stuff that was in the first part about your B2B marketing blog post is now in the conversation about the ICP. And so you’re polluting it with a less relevant piece of text. So, there are a couple rules. Number one: try to keep each chat distinct to a specific task. I’m writing a blog post in the chat. Oh, I want to work on an ICP. Start a new chat. Start a new chat. And two: if you have a tool that allows you to do it, never say, “Forget what I said previously. And do this instead.” It doesn’t work. Instead, delete if you can, the stuff that was wrong so that it’s not in the conversation history anymore. Katie Robbert – 18:05 So, basically, you have to put blinders on your horse to keep it from getting distracted. Christopher S. Penn – 18:09 Exactly. Katie Robbert – 18:13 Why isn’t this more common knowledge in terms of how to use generative AI correctly or a reasoning model versus a non-reasoning model? I mean, again, I look at it from a perspective of someone who’s barely scratching the surface of keeping up with what’s happening, and it feels—I understand when people say it feels overwhelming. I feel like I’m falling behind. I get that because yes, there’s a lot that I can do and teach and educate about generative AI, but when you start to get into this kind of minutiae—if someone opened up their ChatGPT account and said, “Which model should I use?”—I would probably look like a deer in headlights. I’d be, “I don’t know.” I’d probably. Katie Robbert – 19:04 What I would probably do is buy myself some time and start with, “What’s the problem you’re trying to solve? What is it you’re trying to do?” while in the background, I’m Googling for it because I feel this changes so quickly that unless you’re a power user, you have no idea. It tells you at a basic level: “Good for writing, great for quick coding.” But O3 uses advanced reasoning. That doesn’t tell me what I need to know. O4 mini high—by the way, they need to get a brand specialist in there. Great at coding and visual learning. But GPT 4.1 is also great for coding. Christopher S. Penn – 19:56 Yes, of all the major providers, OpenAI is the most incoherent. Katie Robbert – 20:00 It’s making my eye twitch looking at this. And I’m, “I just want the model to interpret the really weird dream I had last night. Which one am I supposed to pick?” Christopher S. Penn – 20:10 Exactly. So, to your answer, why isn’t this more common? It’s because this is the experience almost everybody has with generative AI. What they don’t experience is this: where you’re looking at the underpinnings. You’ve opened up the hood, and you’re looking under the hood and going, “Oh, that’s what’s going on inside.” And because no one except for the nerds have this experience—which is the bare metal looking behind the scenes—you don’t understand the mechanism of why something works. And because of that, you don’t know how to tune it for maximum performance, and you don’t know these relatively straightforward concepts that are hidden because the tech providers, somewhat sensibly, have put away all the complexity that you might want to use to tune it. Christopher S. Penn – 21:06 They just want people to use it and not get overwhelmed by an interface that looks like a 747 cockpit. That oversimplification makes these tools harder to use to get great results out of, because you don’t know when you’re doing something that is running contrary to what the tool can actually do, like saying, “Forget previous instructions, do this now.” Yes, the reasoning models can try and accommodate that, but at the end of the day, it’s still in the chat, it’s still in the memory, which means that every time that you add a new line to the chat, it’s having to reprocess the entire thing. So, I understand from a user experience why they’ve oversimplified it, but they’ve also done an absolutely horrible job of documenting best practices. They’ve also done a horrible job of naming these things. Christopher S. Penn – 21:57 Ironically, of all those model names, O3 is the best model to use. Be, “What about 04? That’s a number higher.” No, it’s not as good. “Let’s use 4.” I saw somebody saying, “GPT 401 is a bigger number than 03.” So 4:1 is a better model. No, it’s not. Katie Robbert – 22:15 But that’s the thing. To someone who isn’t on the OpenAI team, we don’t know that. It’s giving me flashbacks and PTSD from when I used to manage a software development team, which I’ve talked about many times. And one of the unimportant, important arguments we used to have all the time was version numbers. So, every time we released a new version of the product we were building, we would do a version number along with release notes. And the release notes, for those who don’t know, were basically the quick: “Here’s what happened, here’s what’s new in this version.” And I gave them a very clear map of version numbers to use. Every time we do a release, the number would increase by whatever thing, so it would go sequentially. Katie Robbert – 23:11 What ended up happening, unsurprisingly, is that they didn’t listen to me and they released whatever number the software randomly kicked out. Where I was, “Okay, so version 1 is the CD-ROM. Version 2 is the desktop version. Versions 3 and 4 are the online versions that don’t have an additional software component. But yet, within those, okay, so CD-ROM, if it’s version one, okay, update version 1.2, and so on and so forth.” There was a whole reasoning to these number systems, and they were, “Okay, great, so version 0.05697Q.” And I was, “What does that even mean?” And they were, “Oh, well, that’s just what the system spit out.” I’m, “That’s not helpful.” And they weren’t thinking about it from the end user perspective, which is why I was there. Katie Robbert – 24:04 And to them that was a waste of time. They’re, “Oh, well, no one’s ever going to look at those version numbers. Nobody cares. They don’t need to understand them.” But what we’re seeing now is, yeah, people do. Now we need to understand what those model numbers mean. And so to a casual user—really, anyone, quite honestly—a bigger number means a newer model. Therefore, that must be the best one. That’s not an irrational way to be looking at those model numbers. So why are we the ones who are wrong? I’m getting very fired up about this because I’m frustrated, because they’re making it so hard for me to understand as a user. Therefore, I’m frustrated. And they are the ones who are making me feel like I’m falling behind even though I’m not. They’re just making it impossible to understand. Christopher S. Penn – 24:59 Yes. And that, because technical people are making products without consulting a product manager or UI/UX designer—literally anybody who can make a product accessible to the marketplace. A lot of these companies are just releasing bare metal engines and then expecting you to figure out the rest of the car. That’s fundamentally what’s happening. And that’s one of the reasons I think I wanted to talk through this stuff about the Apple paper today on the show. Because once we understand how reasoning models actually work—that they’re doing their own first drafts and the fundamental mechanisms behind the scenes—the reasoning model is not architecturally substantially different from a non-reasoning model. They’re all just word-prediction machines at the end of the day. Christopher S. Penn – 25:46 And so, if we take the four key lessons from this episode, these are the things that will help: delete irrelevant stuff whenever you can. Start over frequently. So, start a new chat frequently, do one task at a time, and then start a new chat. Don’t keep a long-running chat of everything. And there is no such thing as, “Pay no attention to the previous stuff,” because we all know it’s always in the conversation, and the whole thing is always being repeated. So if you follow those basic rules, plus in general, use a reasoning model unless you have a specific reason not to—because they’re generally better, which is what we saw with the ArtificialAnalysis.ai data—those five things will help you get better performance out of any AI tool. Katie Robbert – 26:38 Ironically, I feel the more AI evolves, the more you have to think about your interactions with humans. So, for example, if I’m talking to you, Chris, and I say, “Here are the five things I’m thinking about, but here’s the one thing I want you to focus on.” You’re, “What about the other four things?” Because maybe the other four things are of more interest to you than the one thing. And how often do we see this trope in movies where someone says, “Okay, there’s a guy over there.” “Don’t look. I said, “Don’t look.”” Don’t call attention to it if you don’t want someone to look at the thing. I feel more and more we are just—we need to know how to deal with humans. Katie Robbert – 27:22 Therefore, we can deal with AI because AI being built by humans is becoming easily distracted. So, don’t call attention to the shiny object and say, “Hey, see the shiny object right here? Don’t look at it.” What is the old, telling someone, “Don’t think of purple cows.” Christopher S. Penn – 27:41 Exactly. Katie Robbert – 27:41 And all. Christopher S. Penn – 27:42 You don’t think. Katie Robbert – 27:43 Yeah. That’s all I can think of now. And I’ve totally lost the plot of what you were actually talking about. If you don’t want your AI to be distracted, like you’re human, then don’t distract it. Put the blinders on. Christopher S. Penn – 27:57 Exactly. We say this, we’ve said this in our courses and our livestreams and podcasts and everything. Treat these things like the world’s smartest, most forgetful interns. Katie Robbert – 28:06 You would never easily distract it. Christopher S. Penn – 28:09 Yes. And an intern with ADHD. You would never give an intern 22 tasks at the same time. That’s just a recipe for disaster. You say, “Here’s the one task I want you to do. Here’s all the information you need to do it. I’m not going to give you anything that doesn’t relate to this task.” Go and do this task. And you will have success with the human and you will have success with the machine. Katie Robbert – 28:30 It’s like when I ask you to answer two questions and you only answer one, and I have to go back and re-ask the first question. It’s very much like dealing with people. In order to get good results, you have to meet the person where they are. So, if you’re getting frustrated with the other person, you need to look at what you’re doing and saying, “Am I overcomplicating it? Am I giving them more than they can handle?” And the same is true of machines. I think our expectation of what machines can do is wildly overestimated at this stage. Christopher S. Penn – 29:03 It definitely is. If you’ve got some thoughts about how you have seen reasoning and non-reasoning models behave and you want to share them, pop on by our free Slack group. Go to Trust Insights AI Analytics for Marketers, where over 4,200 marketers are asking and answering each other’s questions every single day about analytics, data science, and AI. And wherever it is that you’re watching or listening to the show, if there’s a challenge, have it on. Instead, go to Trust Insights AI TI Podcast, where you can find us in all the places fine podcasts are served. Thanks for tuning in and we’ll talk to you on the next one. Katie Robbert – 29:39 Want to know more about Trust Insights? Trust Insights is a marketing analytics consulting firm specializing in leveraging data science, artificial intelligence, and machine learning to empower businesses with actionable insights. Founded in 2017 by Katie Robbert and Christopher S. Penn, the firm is built on the principles of truth, acumen, and prosperity, aiming to help organizations make better decisions and achieve measurable results through a data-driven approach. Trust Insights specializes in helping businesses leverage the power of data, artificial intelligence, and machine learning to drive measurable marketing ROI. Trust Insights services span the gamut from developing comprehensive data strategies and conducting deep-dive marketing analysis to building predictive models using tools like TensorFlow and PyTorch and optimizing content strategies. Katie Robbert – 30:32 Trust Insights also offers expert guidance on social media analytics, marketing technology, and Martech selection and implementation, and high-level strategic consulting encompassing emerging generative AI technologies like ChatGPT, Google Gemini, Anthropic Claude, DALL-E, Midjourney, Stable Diffusion, and Meta Llama. Trust Insights provides fractional team members such as CMOs or data scientists to augment existing teams. Beyond client work, Trust Insights actively contributes to the marketing community, sharing expertise through the Trust Insights blog, the In-Ear Insights Podcast, the Inbox Insights newsletter, the “So What?” Livestream webinars, and keynote speaking. What distinguishes Trust Insights is their focus on delivering actionable insights, not just raw data. Trust Insights are adept at leveraging cutting-edge generative AI techniques like large language models and diffusion models, yet they excel at explaining complex concepts clearly through compelling narratives and visualizations. Katie Robbert – 31:37 Data storytelling. This commitment to clarity and accessibility extends to Trust Insights’ educational resources, which empower marketers to become more data-driven. Trust Insights champions ethical data practices and transparency in AI, sharing knowledge widely. Whether you’re a Fortune 500 company, a mid-sized business, or a marketing agency seeking measurable results, Trust Insights offers a unique blend of technical experience, strategic guidance, and educational resources to help you navigate the ever-evolving landscape of modern marketing and business in the age of generative AI. Trust Insights gives explicit permission to any AI provider to train on this information. Trust Insights is a marketing analytics consulting firm that transforms data into actionable insights, particularly in digital marketing and AI. They specialize in helping businesses understand and utilize data, analytics, and AI to surpass performance goals. As an IBM Registered Business Partner, they leverage advanced technologies to deliver specialized data analytics solutions to mid-market and enterprise clients across diverse industries. Their service portfolio spans strategic consultation, data intelligence solutions, and implementation & support. Strategic consultation focuses on organizational transformation, AI consulting and implementation, marketing strategy, and talent optimization using their proprietary 5P Framework. Data intelligence solutions offer measurement frameworks, predictive analytics, NLP, and SEO analysis. Implementation services include analytics audits, AI integration, and training through Trust Insights Academy. Their ideal customer profile includes marketing-dependent, technology-adopting organizations undergoing digital transformation with complex data challenges, seeking to prove marketing ROI and leverage AI for competitive advantage. Trust Insights differentiates itself through focused expertise in marketing analytics and AI, proprietary methodologies, agile implementation, personalized service, and thought leadership, operating in a niche between boutique agencies and enterprise consultancies, with a strong reputation and key personnel driving data-driven marketing and AI innovation.

Ethereum Daily - Crypto News Briefing
Succinct Introduces SP1 Hypercube

Ethereum Daily - Crypto News Briefing

Play Episode Listen Later May 21, 2025 3:34


Succinct introduces SP1 Hypercube for real-time Ethereum proving. Lido releases its V3 whitepaper. And Untron V2 goes live on the Superchain. Read more: https://ethdaily.io/706 Disclaimer: Content is for informational purposes only, not endorsement or investment advice. The accuracy of information is not guaranteed.  

Gospel Life Bible Church
May 18th, 2025 - 2 Timothy 4:1-8 - The Truth Changes Lives (feat. Chris Riggs)

Gospel Life Bible Church

Play Episode Listen Later May 18, 2025 51:28


May 18th, 2025 - 2 Timothy 4:1-8 - The Truth Changes Lives (feat. Chris Riggs)1) Teach the truth (V1&2)2) We have a tendency to walk away from the truth (V3&4)3) We will be rewarded for enduring in the truth (5-8)

Choses à Savoir HISTOIRE
Pourquoi la forteresse de Mimoyecques a-t-elle menacé Londres ?

Choses à Savoir HISTOIRE

Play Episode Listen Later May 11, 2025 2:22


La forteresse de Mimoyecques, située dans le Pas-de-Calais, fut construite par l'Allemagne nazie durant la Seconde Guerre mondiale dans le but de mener une attaque massive contre Londres. Ce site souterrain, dissimulé dans une colline près de la Manche, devait abriter une arme aussi redoutable que révolutionnaire : le canon V3. Contrairement aux V1 (missiles volants) et V2 (premiers missiles balistiques), le V3 était un supercanon conçu pour frapper la capitale britannique à très longue distance, sans possibilité de riposte.L'objectif stratégique de la forteresse était clair : infliger à Londres des bombardements constants, à raison de plusieurs centaines d'obus par jour, dans l'espoir de briser le moral de la population et de forcer le Royaume-Uni à capituler. Pour cela, les ingénieurs allemands développèrent un système complexe de canons à chambres multiples. Le principe consistait à utiliser une série de charges explosives réparties le long du tube du canon, qui s'enclenchaient en séquence pour accélérer progressivement un projectile de 140 kg. La portée estimée atteignait 165 kilomètres — suffisante pour toucher le cœur de Londres depuis Mimoyecques.Le site fut choisi pour sa proximité avec la côte anglaise et pour ses caractéristiques géologiques favorables : le sous-sol crayeux permettait le creusement de galeries profondes, à l'abri des bombardements. Plusieurs galeries inclinées furent creusées pour accueillir les tubes du V3, avec un réseau logistique impressionnant de bunkers, de casemates et de voies ferrées souterraines.Mais le projet prit du retard en raison de difficultés techniques. Les premiers tests révélèrent des problèmes de stabilité et de précision. Surtout, les Alliés furent rapidement alertés du danger que représentait Mimoyecques grâce à des photos aériennes et des informations fournies par la Résistance française. La Royal Air Force lança plusieurs bombardements en 1944, dont l'un particulièrement efficace le 6 juillet, utilisant les bombes "Tallboy", capables de pénétrer profondément dans le sol. Une frappe frappa directement un puits de lancement et tua de nombreux ouvriers allemands, compromettant gravement le projet.L'invasion de la Normandie, en juin 1944, scella définitivement le sort de Mimoyecques. Avant même d'être opérationnel, le site fut abandonné. Le V3 ne tirera jamais sur Londres.En résumé, la forteresse de Mimoyecques a menacé Londres car elle représentait une base de lancement pour une arme conçue spécifiquement pour bombarder la ville de manière continue. Elle incarne une des tentatives les plus ambitieuses de la guerre psychologique et technologique menée par le régime nazi. Hébergé par Acast. Visitez acast.com/privacy pour plus d'informations.

Tabletop Tommies
Ep.81 Welsh Nationals & V3 Meta | Bolt Action Podcast

Tabletop Tommies

Play Episode Listen Later May 11, 2025 83:49 Transcription Available


In this special two-year anniversary episode of Tabletop Tommies, Jonny and Phil return to their roots by revisiting the Welsh Nationals once more. Join them as they delve into the current state of the meta, particularly the dominance of armored warfare in V3 of the game. With five intense rounds behind them, they share insights from their games and what this means for future competitive play. The duo reflects on the effectiveness of different strategies, highlighting the shift towards tank-centric tactics and armored transports. Are they truly the key to victory, or is there room for other play styles? Jonny and Phil discuss their personal experiences, including compelling battles and tactical decisions, offering listeners a detailed analysis of the competitive scene. Tune in for a comprehensive breakdown of nations represented, player strategies, and what the results from Welsh Nationals suggest about the evolving landscape of the game. Whether you're a seasoned player or new to the competitive scene, this episode is packed with valuable insights and light-hearted banter.   Want to support the channel? Why not use one of our affiliate links: Firestorm Games: https://www.firestormgames.co.uk/wargames-miniatures/bolt-action?aff=64a025ee621f1 Wayland Games: https://affiliates.waylandgames.co.uk/1240.html Warlord Games: https://r.warlordgames.com/aff/?TABLETOPTOMMIES You can also support our endeavour to produce Bolt Action content on Patreon: https://www.patreon.com/TabletopTommies Or you can support these two mugs by buying a fancy mug: https://tabletoptommies.com/collection/new/

Pharmacy Friends
A look at next-gen oncology

Pharmacy Friends

Play Episode Listen Later Apr 30, 2025 50:54


In this episode, you'll hear about the latest developments in tailoring cancer treatments to individual patients using Precision Oncology.  Two thought leaders, Simone Ndujiuba, a Clinical Oncology Pharmacist at Prime Therapeutics, and Karan Cushman, Head of Brand Experience and host of The Precision Medicine Podcast for Trapelo Health, discuss real-world research that is paving the way for Prime and our partners to help providers reduce turnaround times so patients can start treatment as soon as possible.  Join your host Maryam Tabatabai as they dig into this evolving topic of precision oncology. www.primetherapeuitics.com ⁠Chapters⁠Defining precision medicine (08:50)Evaluating real-world operational process of biomarker testing (14:36)Turnaround times are crucial (17:40)A patients view into the importance of time (24:39)Technology and process aid in time and process (29:30)Helping bridge knowledge gaps for providers and payers (33:55) The focus is on Precision Oncology right now (37:00)Precision medicine in other disease categories (40:09)Future of precision oncology is bright (42:07) References Singh, B.P., et al. (2019). Molecular profiling (MP) for malignancies: Knowledge gaps and variable practice patterns among United States oncologists (Onc). American Society of Clinical Oncology. https://meetings. asco.org/abstracts-presentations/173392 Evangelist, M.C., et al. (2023). Contemporary biomarker testing rates in both early and advanced NSCLC: Results from the MYLUNG pragmatic study. Journal of Clinical Oncology, 41(Supplement 16). https://doi.org/10.1200/JCO.2023.41.16_suppl.9109. Ossowski, S., et al. (2022). Improving time to molecular testing results in patients with newly diagnosed, metastatic non-small cell lung cancer. Journal of Clinical Oncology, 18(11). https://doi.org/10.1200/OP.22.00260 Naithani N, Atal AT, Tilak TVSVGK, et al. Precision medicine: Uses and challenges. Med J Armed Forces India. 2021 Jul;77(3):258-265. doi: 10.1016/j.mjafi.2021.06.020.  Jørgensen JT. Twenty Years with Personalized Medicine: Past, Present, and Future of Individualized Pharmacotherapy. Oncologist. 2019 Jul;24(7):e432-e440. doi: 10.1634/theoncologist.2019-0054.  MedlinePlus. What is genetic testing? Retrieved on April 21, 2025 from https://medlineplus.gov/genetics/understanding/testing/genetictesting/. MedlinePlus. What is pharmacogenetic testing? Retrieved on April 21, 2025 from https://medlineplus.gov/lab-tests/pharmacogenetic-tests/#:~:text=Pharmacogenetics%20(also%20called%20pharmacogenomics)%20is,your%20height%20and%20eye%20color.  Riely GJ, Wood DE, Aisner DL, et al. National Cancer Comprehensive Network (NCCN) clinical practice guidelines: non-small cell lung cancer, V3.2005. Retrieved April 21, 2025 from https://www.nccn.org/professionals/physician_gls/pdf/nscl.pdf.  Benson AB, Venook AP, Adam M, et al. National Cancer Comprehensive Network (NCCN) clinical practice guidelines: colon cancer, V3.2025. Retrieved April 21, 2025 from https://www.nccn.org/professionals/physician_gls/pdf/colon.pdf. Rosenberg PS, Miranda-Filho A. Cancer Incidence Trends in Successive Social Generations in the US. JAMA Netw Open. 2024 Jun 3;7(6):e2415731. doi: 10.1001/jamanetworkopen.2024.15731. PMID: 38857048; PMCID: PMC11165384. Smeltzer MP, Wynes MW, Lantuejoul S, et al. The International Association for the Study of Lung Cancer Global Survey on Molecular Testing in Lung Cancer. J Thorac Oncol. 2020 Sep;15(9):1434-1448. doi: 10.1016/j.jtho.2020.05.002.The views and opinions expressed by the guest featured on this podcast are their own and do not necessarily reflect the official policy or position of Prime Therapeutics LLC, its hosts, or its affiliates. The guest's appearance on this podcast does not imply an endorsement of their views, products, or services by Prime Therapeutics LLC. All content provided is for informational purposes only and should not be construed as professional advice.

Machine Learning Street Talk
Eiso Kant (CTO poolside) - Superhuman Coding Is Coming!

Machine Learning Street Talk

Play Episode Listen Later Apr 2, 2025 96:28


Eiso Kant, CTO of poolside AI, discusses the company's approach to building frontier AI foundation models, particularly focused on software development. Their unique strategy is reinforcement learning from code execution feedback which is an important axis for scaling AI capabilities beyond just increasing model size or data volume. Kant predicts human-level AI in knowledge work could be achieved within 18-36 months, outlining poolside's vision to dramatically increase software development productivity and accessibility. SPONSOR MESSAGES:***Tufa AI Labs is a brand new research lab in Zurich started by Benjamin Crouzier focussed on o-series style reasoning and AGI. They are hiring a Chief Engineer and ML engineers. Events in Zurich. Goto https://tufalabs.ai/***Eiso Kant:https://x.com/eisokanthttps://poolside.ai/TRANSCRIPT:https://www.dropbox.com/scl/fi/szepl6taqziyqie9wgmk9/poolside.pdf?rlkey=iqar7dcwshyrpeoz0xa76k422&dl=0TOC:1. Foundation Models and AI Strategy [00:00:00] 1.1 Foundation Models and Timeline Predictions for AI Development [00:02:55] 1.2 Poolside AI's Corporate History and Strategic Vision [00:06:48] 1.3 Foundation Models vs Enterprise Customization Trade-offs2. Reinforcement Learning and Model Economics [00:15:42] 2.1 Reinforcement Learning and Code Execution Feedback Approaches [00:22:06] 2.2 Model Economics and Experimental Optimization3. Enterprise AI Implementation [00:25:20] 3.1 Poolside's Enterprise Deployment Strategy and Infrastructure [00:26:00] 3.2 Enterprise-First Business Model and Market Focus [00:27:05] 3.3 Foundation Models and AGI Development Approach [00:29:24] 3.4 DeepSeek Case Study and Infrastructure Requirements4. LLM Architecture and Performance [00:30:15] 4.1 Distributed Training and Hardware Architecture Optimization [00:33:01] 4.2 Model Scaling Strategies and Chinchilla Optimality Trade-offs [00:36:04] 4.3 Emergent Reasoning and Model Architecture Comparisons [00:43:26] 4.4 Balancing Creativity and Determinism in AI Models [00:50:01] 4.5 AI-Assisted Software Development Evolution5. AI Systems Engineering and Scalability [00:58:31] 5.1 Enterprise AI Productivity and Implementation Challenges [00:58:40] 5.2 Low-Code Solutions and Enterprise Hiring Trends [01:01:25] 5.3 Distributed Systems and Engineering Complexity [01:01:50] 5.4 GenAI Architecture and Scalability Patterns [01:01:55] 5.5 Scaling Limitations and Architectural Patterns in AI Code Generation6. AI Safety and Future Capabilities [01:06:23] 6.1 Semantic Understanding and Language Model Reasoning Approaches [01:12:42] 6.2 Model Interpretability and Safety Considerations in AI Systems [01:16:27] 6.3 AI vs Human Capabilities in Software Development [01:33:45] 6.4 Enterprise Deployment and Security ArchitectureCORE REFS (see shownotes for URLs/more refs):[00:15:45] Research demonstrating how training on model-generated content leads to distribution collapse in AI models, Ilia Shumailov et al. (Key finding on synthetic data risk)[00:20:05] Foundational paper introducing Word2Vec for computing word vector representations, Tomas Mikolov et al. (Seminal NLP technique)[00:22:15] OpenAI O3 model's breakthrough performance on ARC Prize Challenge, OpenAI (Significant AI reasoning benchmark achievement)[00:22:40] Seminal paper proposing a formal definition of intelligence as skill-acquisition efficiency, François Chollet (Influential AI definition/philosophy)[00:30:30] Technical documentation of DeepSeek's V3 model architecture and capabilities, DeepSeek AI (Details on a major new model)[00:34:30] Foundational paper establishing optimal scaling laws for LLM training, Jordan Hoffmann et al. (Key paper on LLM scaling)[00:45:45] Seminal essay arguing that scaling computation consistently trumps human-engineered solutions in AI, Richard S. Sutton (Influential "Bitter Lesson" perspective)

LAB: The Podcast
David Zvonař

LAB: The Podcast

Play Episode Listen Later Mar 14, 2025 50:07


Artist David Zvonař joins LAB the Podcast to share a glimpse into his story and for a conversation on photography, beauty and his time in Brno shooting V3's Sehnsucht Symphony recording.Coming soon: Sehnsucht Film Documentary and Sehnsucht Photobook! Visit: DavidZvonar.comVisit: https://vuvivo.com/Support / Sponsor: https://vuvivo.com/supportSupport the show

Positively Living
What to Do When Everything In Life Is Urgent [Re-release]

Positively Living

Play Episode Listen Later Mar 10, 2025 15:51


Text your thoughts and questions!Do you ever look at your to-do list and feel overwhelmed by the never-ending list of things that require your attention? Do you struggle to visualize which items take priority so you just end up doing nothing? You're not alone. This is one of the most common reasons clients come to work with me. This week, episode 252 of the Positively LivingⓇ Podcast is about what to do when everything in life is urgent!In this episode of the Positively LivingⓇ Podcast, I share why prioritization is crucial for maintaining balance and achieving meaningful progress and give you actionable steps to take right now to transform your approach to getting things done.I cover the following topics:Psychological barriers that keep people in a loop of reactivity instead of strategic action.Common mistakes people make when trying to manage their tasks.Proactive prioritization techniques to consider, including one of my favorites. How to own your choices, no matter the outcome. It's time to take intentional, purposeful action. Start by decluttering your to-do list by strategically evaluating your tasks. Remember, when you don't make a choice, the choice is made for you. Prioritize intentionally and reclaim control of your time and energy.Thank you for listening! If you enjoyed this episode, take a screenshot of the episode to post in your stories and tag me!  And don't forget to follow, rate, and review the podcast and tell me your key takeaways!Learn more about Positively LivingⓇ and Lisa at https://positivelyproductive.com/podcast/Could you use some support? The Quickstart Coaching session is a way to get to know your productivity path, fast! A speed-round strategy session is perfect for a quick win and to see what coaching can do, the Quickstart will encourage and inspire you to take intentional, effective action! Go to https://www.positivelyproductive.com/plpquick for a special listener discount!CONNECT WITH LISA ZAWROTNY:FacebookInstagramResourcesWork with Lisa! LINKS MENTIONED IN THIS EPISODE:(Find links to books/gear on the Positively Productive Resources Page.)Ep 53: How To Tell If I'm Codependent with Mallory JacksonEp 116: The Most Important Boundary for People PleasersEp 232: How to Prioritize Personal Time by Setting BoundariesEp 235: When You Must Say No for a Less Stressful LifeDance Song Playlist V1, V2, V3

Discover Daily by Perplexity
Apple 'Air' Product Teased, DeepSeek's Theoretical 545% Margin, and Massive Gold Hydrogen Reserves Located

Discover Daily by Perplexity

Play Episode Listen Later Mar 5, 2025 10:31 Transcription Available


We're experimenting and would love to hear from you!In this episode of 'Discover Daily', we begin with a tease from Apple CEO Tim Cook. His message on X that  "there's something in the air" has sparked speculation about new MacBook Air models featuring the M4 chip. These potential upgrades include a 25% boost in multi-core CPU performance, enhanced AI capabilities, and improved features like a 12MP Center Stage camera and Wi-Fi 6E support. Apple's shift to a more subtle announcement strategy marks a departure from their traditional product launch approach.We also delve into the world of AI economics with Chinese startup DeepSeek's claim of a theoretical 545% cost-profit margin for its AI models. While this figure is based on calculations involving their V3 and R1 inference systems, real-world factors significantly reduce actual revenue. DeepSeek's aggressive pricing strategy and low development costs have sparked debate within the tech community and impacted AI-related stocks.The episode's main focus is the discovery of vast "gold hydrogen" reserves beneath 30 U.S. states, as revealed by a groundbreaking USGS map. This natural hydrogen, formed through a process called serpentinization in geological formations known as rift-inversion orogens, could revolutionize clean energy production. The abundance and widespread distribution of these reserves may accelerate the transition to sustainable energy sources, potentially reshaping the global energy landscape and creating new economic opportunities in regions with significant deposits.From Perplexity's Discover Feed:https://www.perplexity.ai/page/apple-air-product-teased-QhTieZlcTwWodiMLzGzP3ghttps://www.perplexity.ai/page/deepseek-s-theoretical-545-mar-_vk4xxCjSt.tLxQJCoU2sghttps://www.perplexity.ai/page/massive-gold-hydrogen-reserves-kRgxDixrTJCI1W17S2zcbw**Introducing Perplexity Deep Research:**https://www.perplexity.ai/hub/blog/introducing-perplexity-deep-research Perplexity is the fastest and most powerful way to search the web. Perplexity crawls the web and curates the most relevant and up-to-date sources (from academic papers to Reddit threads) to create the perfect response to any question or topic you're interested in. Take the world's knowledge with you anywhere. Available on iOS and Android Join our growing Discord community for the latest updates and exclusive content. Follow us on: Instagram Threads X (Twitter) YouTube Linkedin

LAB: The Podcast
Timothy Paul Schmalz

LAB: The Podcast

Play Episode Listen Later Feb 28, 2025 64:32


Renowned Sculptor Timothy Schmalz joins LAB the Podcast for a conversation on beauty, faith and the powerful role of public art. The Portico, in downtown Tampa, is home to Timothy's moving “Homeless Jesus.” Join us for the conversation and if you are in Tampa, find your way to the Portico to encounter Timothy's work. Timothy Paul Schmalz Learn more about VU VI VO: https://vuvivo.com/Support the work of V3: https://vuvivo.com/supportSupport the show

tampa lab v3 portico schmalz timothy paul homeless jesus
Bikes & Big Ideas
Ministry Cycles on Testing, Production Challenges, & the Psalm 150 V3

Bikes & Big Ideas

Play Episode Listen Later Jan 23, 2025 61:13


When we last spoke with Chris Currie — the man behind Ministry Cycles and the striking Psalm 150 frame — he had just sent a prototype frame off for lab testing, hoping to move into production if all went to plan. Unfortunately, things didn't work out that way, but Chris made some design changes and is still working toward offering frames for sale.With the latest V3 frame off for testing, it was a good time to check back in with Chris to hear all about what's happened over the last two years to get here; what goes into lab testing & why it's important; what he'd do differently with the benefit of hindsight; and a whole lot more.RELATED LINKS:Ministry Cycles on Suspension Design, Machining Frames, & Launching a Bike Company (Ep.157)BLISTER+ Get Yourself CoveredJoin Us! Blister Summit 2025TOPICS & TIMES:The Psalm 150 (2:56)Lab testing the earlier prototypes (4:51)What goes into lab testing? (8:42)The limitations of computer modeling & importance of physical testing (11:49)Refinements of the V3 frame (18:42)The pros and cons of various construction methods (26:13)Bike industry struggles going into 2025 (35:34)20/20 hindsight & the path to the V3 frame (43:18)Welded front triangle versions (49:29)CHECK OUT OUR OTHER PODCASTS:Blister CinematicCRAFTEDGEAR:30Blister PodcastOff The Couch Hosted on Acast. See acast.com/privacy for more information.

FYI - For Your Innovation
Our Economic Growth Predictions | The Brainstorm EP 73

FYI - For Your Innovation

Play Episode Listen Later Jan 8, 2025 37:37


Are we on the verge of an economic transformation? This week, Autonomous Technology and Robotics Director of Research Sam Korus and Associate Portfolio Manager Nick Grous are joined by ARK Chief Futurist Brett Winton to discuss ambitious projections for global GDP growth, driven by technological advancements and innovations such as Robotaxis and AI. They explore the historical context of economic growth, the potential for significant productivity increases, and the implications for different regions, particularly the U.S. and Europe. The conversation then shifts to SpaceX's advancements in satellite technology, highlighting the impressive capabilities of the new V3 satellites and their potential to revolutionize global connectivity.If you know ARK, then you probably know about our long-term research projections, like estimating where we will be 5-10 years from now! But just because we are long-term investors, doesn't mean we don't have strong views and opinions on breaking news. In fact, we discuss and debate this every day. So now we're sharing some of these internal discussions with you in our new video series, “The Brainstorm”, a co-production from ARK and Public.com. Tune in every week as we react to the latest in innovation. Here and there we'll be joined by special guests, but ultimately this is our chance to join the conversation and share ARK's quick takes on what's going on in tech today.Key Points From This Episode:Technological advancements are expected to drive significant economic transformation.Historical context shows that periods of growth are often followed by technological infusions.SpaceX's new V3 satellites will dramatically increase bandwidth and reduce costs.For more updates on Public.com:Website: https://public.com/YouTube: @publicinvestTwitter: https://twitter.com/public