Domesticated species of South American camelid
POPULARITY
Categories
Safety Sheriff Labrador|Safety Story for Kids|Safety Tips|BabyBus
Ladies, today we're talking all about haircuts from hell. And speaking of the afterlife, Kristin has a very special hot flash: Pope Leo. That's right, our Chicago Pope. The Lord Be with you (and with da Bears).If you enjoyed this episode, leave a review and make sure you SUBSCRIBE!If you are interested in advertising on this podcast email ussales@acast.comTo request #IMOMSOHARD to be on your Podcast, Radio Show, or TV Show, reach out to talent@pionairepodcasting.comGet our sponsor DISCOUNT CODES here! https://linktr.ee/imshpodcastFOLLOW US:IG: instagram.com/imomsohardYouTube: youtube.com/imomsohard Hosted on Acast. See acast.com/privacy for more information.
Court will decide if supermarket discriminated by not allowing alpaca in store. Kansas Mom Sues Porn Sites Because Her 14 Year Old Son Visited Some. Psychopaths Are More Attractive, Study Warns. // SUPPORT by joining the Weird AF News Patreon http://patreon.com/weirdafnews - OR buy Jonesy a coffee at http://buymeacoffee.com/funnyjones Buy MERCH: https://weirdafnews.merchmake.com/ - Check out the official website https://WeirdAFnews.com and FOLLOW host Jonesy at http://instagram.com/funnyjones
On this episode of Bounced From The Roadhouse:Special Guests in 4B: Leslie Lane and Jeanie with the Artemis HouseCoffee TalkRandom FactsAirbnbAlpaca PurposeMorris the CrocodileKim Kardashian JewelryApple iPhone LawsuitBowling Rings BackCat QuinceañeraBurned 100 BooksCost Per Live SongPsychopaths are Better LookingArtemis House InterviewQuestions? Comments? Leave us a message! 605-343-6161Don't forget to subscribe, leave us a review and some stars Hosted on Acast. See acast.com/privacy for more information.
Welcome to the podcast for alpaca people!In this episode, I consider the difficulty of doing some alpaca jobs alone.There is also a valley update about the new ready-made family of geese joining the number one family.Thanks for listening and I hope you enjoyed it.If you would like to be in touch, please contact me by email - steve@alpacatribe.com - or leave me a voicemail from your browser.Alpaca Tribe is hosted and produced by Steve Heatherington of Good Podcasting Works, which is part of The Waterside (Swansea) LtdThis podcast uses the following third-party services for analysis: OP3 - https://op3.dev/privacy
The Alpaca in the Room is 1 year old! Woop Woop. To celebrate our 1 year birthday, it's only right that we go back to our origins! Episode 1 was 'what is wrong with linkedin'. So today we ask asking ourselves 'How is Linkedin in 2025?'How is activity on Linkedin affected by other social networks? Are there any topics on Linkedin that don't reference AI? Can we all agree everyone is on linkedin to take, rather than give? All this and we debate if Stephen Hawking was eaten by a shark! Enjoy and thank you for being with us this year.
David Nathan vertelt over optreden met alpaca's, crewleden met slaapproblemen, vergeten illusies, gescheurde enkelbanden en wat als de koffer met je hele show niet in het vliegtuig zit?David Nathan is goochelaar en illusionist. Hij treed op bij beurzen, evenementen, wintercircussen en op cruiseschepen. Daarnaast is hij magisch consultant van onder andere Hans Klok en Victor Mids. Meer informatie over David kun je vinden op www.davidnathan.nlDood Konijn Podcast:Deze podcast wordt gemaakt door Tim Horsting en Janse Heijn. Twee prijswinnende goochelaars uit Wageningen, samen delen zij een passie voor goochelkunst, theater én bellen zij elkaar altijd op na een rampzalig optreden. Geïnspireerd op de "elektra podcast" maken zij samen Dood Konijn.Meer informatie kun je vinden op www.bedrijfsgoochelaar.nl voor Janse en www.timhorsting.nl voor Tim
Bienvenidos a un nuevo episodio de Spicy4tuna. En el día de hoy hablaremos de la suerte en los negocios, las cuentas millonarias de la Iglesia, el negocio de la tela de Alpaca, los avances en inteligencia artificial, la experiencia de Willyrex grabándose un día entero, y mucho más. Así que sin más dilación, empecemos. Crea tu Página Web con Hostinger: https://www.hostinger.com/spicy4tuna ️ Cupón de 10% de Descuento para planes de +12 meses: SPICY4TUNA ️ Consigue tus entradas para Madrid Economic Forum con un 10% de descuento con el cupón SPICY: https://madrid.economicforum.es/?referral-code=spicy4tuna ️ Disfruta de 30 días gratis y acceder a los mejores podcast sin anuncios en Podimo: https://go.podimo.com/spicy4tuna ☕ Prueba el mejor café de especialidad directo a la puerta de tu casa con Incapto: https://bit.ly/4aicFHu Contacta con Hausum para pedir más información sobre su servicio de inspección de viviendas: https://spicy4tuna.com/hausum Invierte en inmuebles de forma pasiva y sin dolores de cabeza con Invesiva: https://link.inversiva.com/spicy4tuna_youtube Encuentra tu hogar con un alquiler con opción a compra fácil y flexible con Wannaprop: https://wannaprop.es/?utm_source=youtube&utm_medium=spicy4tuna&utm_campaign=acceso_a_la_vivienda : Invierte de forma segura y recibe un 2,27% sobre tu efectivo con Trade Republic: https://trade.re/spicy4tuna Invertir conlleva riesgos, los rendimientos no están garantizados. Aplican T&Cs. Abre tu cuenta de empresa en Finom y comienza a operar en 24h: https://bit.ly/SpicyFinom Aprende a hablar inglés como un Nativo: https://youtalkonline.com/spicy4tuna ️ El curso digital #1 de Oratoria y Comunicación para Hablar en Público con Confianza: https://go.hotmart.com/L97199651U ════════════════ ️ Accede a la Web de Spicy4tuna y Suscríbete a nuestra Newsletter: https://www.spicy4tuna.com Contacto para Sponsors ➡ https://tally.so/r/nrPNE5 Email de Contacto ➡ podcast@spicy4tuna.com ════════════════ Todos los episodios completos: https://www.youtube.com/playlist?list=PL9XxulgDZKuzf6zuPWcuF6anvQOrukMom ════════════════ REDES SOCIALES DE SPICY4TUNA ➜ INSTAGRAM: https://www.instagram.com/spicy4tunapodcast/ ➜ TIKTOK: https://www.tiktok.com/@spicy4tuna ➜ FACEBOOK: https://www.facebook.com/spicy4tuna ════════════════ ️ ESCUCHA SPICY4TUNA EN FORMATO PODCAST Spotify: https://open.spotify.com/show/2QPC17Z9LhTntCA4c3Ijk9?si=39b610a14bb24f1f iTunes: https://podcasts.apple.com/es/podcast/spicy4tuna/id1714279648 iVoox: https://www.ivoox.com/escuchar-audios-spicy4tuna_al_33258956_1.html ════════════════ ¿QUIÉNES SOMOS? · Euge Oller: https://www.instagram.com/euge.oller/ · Willyrex: https://www.instagram.com/willyrex/ · Marc Urgell: https://www.instagram.com/marcurgelldiaz/ · Alvaro845: https://www.instagram.com/alvaro845/ ════════════════ CAPÍTULOS 00:00:00 INTRODUCCIÓN 00:04:18 SUERTE Y NEGOCIOS 00:27:52 EL NEGOCIO DE LA IGLESIA 00:53:06 AVANZES EN INTELIGENCIA ARTIFICIAL 01:19:51 EL NEGOCIO DE LA TELA DE ALPACA 01:29:12 MADRID ECONOMIC FORUM 01:39:57 CURSO GRATIS DE EUGE OLLER 01:43:38 UN DÍA CON WILLYREX 01:50:07 FILMS & BUSINESS
Bienvenidos a un nuevo episodio de Spicy4tuna. En el día de hoy hablaremos de la suerte en los negocios, las cuentas millonarias de la Iglesia, el negocio de la tela de Alpaca, los avances en inteligencia artificial, la experiencia de Willyrex grabándose un día entero, y mucho más. Así que sin más dilación, empecemos. Crea tu Página Web con Hostinger: https://www.hostinger.com/spicy4tuna ️ Cupón de 10% de Descuento para planes de +12 meses: SPICY4TUNA ️ Consigue tus entradas para Madrid Economic Forum con un 10% de descuento con el cupón SPICY: https://madrid.economicforum.es/?referral-code=spicy4tuna ️ Disfruta de 30 días gratis y acceder a los mejores podcast sin anuncios en Podimo: https://go.podimo.com/spicy4tuna ☕ Prueba el mejor café de especialidad directo a la puerta de tu casa con Incapto: https://bit.ly/4aicFHu Contacta con Hausum para pedir más información sobre su servicio de inspección de viviendas: https://spicy4tuna.com/hausum Invierte en inmuebles de forma pasiva y sin dolores de cabeza con Invesiva: https://link.inversiva.com/spicy4tuna_youtube Encuentra tu hogar con un alquiler con opción a compra fácil y flexible con Wannaprop: https://wannaprop.es/?utm_source=youtube&utm_medium=spicy4tuna&utm_campaign=acceso_a_la_vivienda : Invierte de forma segura y recibe un 2,27% sobre tu efectivo con Trade Republic: https://trade.re/spicy4tuna Invertir conlleva riesgos, los rendimientos no están garantizados. Aplican T&Cs. Abre tu cuenta de empresa en Finom y comienza a operar en 24h: https://bit.ly/SpicyFinom Aprende a hablar inglés como un Nativo: https://youtalkonline.com/spicy4tuna ️ El curso digital #1 de Oratoria y Comunicación para Hablar en Público con Confianza: https://go.hotmart.com/L97199651U ════════════════ ️ Accede a la Web de Spicy4tuna y Suscríbete a nuestra Newsletter: https://www.spicy4tuna.com Contacto para Sponsors ➡ https://tally.so/r/nrPNE5 Email de Contacto ➡ podcast@spicy4tuna.com ════════════════ Todos los episodios completos: https://www.youtube.com/playlist?list=PL9XxulgDZKuzf6zuPWcuF6anvQOrukMom ════════════════ REDES SOCIALES DE SPICY4TUNA ➜ INSTAGRAM: https://www.instagram.com/spicy4tunapodcast/ ➜ TIKTOK: https://www.tiktok.com/@spicy4tuna ➜ FACEBOOK: https://www.facebook.com/spicy4tuna ════════════════ ️ ESCUCHA SPICY4TUNA EN FORMATO PODCAST Spotify: https://open.spotify.com/show/2QPC17Z9LhTntCA4c3Ijk9?si=39b610a14bb24f1f iTunes: https://podcasts.apple.com/es/podcast/spicy4tuna/id1714279648 iVoox: https://www.ivoox.com/escuchar-audios-spicy4tuna_al_33258956_1.html ════════════════ ¿QUIÉNES SOMOS? · Euge Oller: https://www.instagram.com/euge.oller/ · Willyrex: https://www.instagram.com/willyrex/ · Marc Urgell: https://www.instagram.com/marcurgelldiaz/ · Alvaro845: https://www.instagram.com/alvaro845/ ════════════════ CAPÍTULOS 00:00:00 INTRODUCCIÓN 00:04:18 SUERTE Y NEGOCIOS 00:27:52 EL NEGOCIO DE LA IGLESIA 00:53:06 AVANZES EN INTELIGENCIA ARTIFICIAL 01:19:51 EL NEGOCIO DE LA TELA DE ALPACA 01:29:12 MADRID ECONOMIC FORUM 01:39:57 CURSO GRATIS DE EUGE OLLER 01:43:38 UN DÍA CON WILLYREX 01:50:07 FILMS & BUSINESS
To celebrate our 500th episode, we bring you the only TPK recorded entirely in person! It’s an extra-long session recorded around a table, complete with the sounds of dice rolling, soda-pop cans opening, and chairs scraping the floor as people move around. The only way to feel more like you’re there is to watch the 360° version on a VR headset! This episode kicks off another Candlekeep mystery in which N’iss, Naiasa, Valenthe, Merdri, and Quewaris investigate a trembling tower. Tony Sindelar with Annette Wierstra, Dan Moren, James Thomson, Jason Snell and Nick Scott.
To celebrate our 500th episode, we bring you the only TPK recorded entirely in person! It’s an extra-long session recorded around a table, complete with the sounds of dice rolling, soda-pop cans opening, and chairs scraping the floor as people move around. The only way to feel more like you’re there is to watch the 360° version on a VR headset! This episode kicks off another Candlekeep mystery in which N’iss, Naiasa, Valenthe, Merdri, and Quewaris investigate a trembling tower. Tony Sindelar with Annette Wierstra, Dan Moren, James Thomson, Jason Snell and Nick Scott.
Dios no busca perfección, busca disposición: Él puede convertir una relación rota en una obra de arte. Mensaje del 04 de mayo del 2025 en el campus Surco de la iglesia Caminodevida - Lima, Perú por los pastores Victor & Rossana Alpaca.
Alleged Manipulation of #Alpaca (BSC) Causes Major Pump-and-Dump (OOC) #Crypto #Cryptocurrency #podcast #BasicCryptonomics Website: https://www.CryptoTalkRadio.net Facebook: @ThisIsCTR Discord: @CryptoTalkRadio Chapters (00:00:00) - Out of Cycle Update(00:00:38) - Alpaca: Was This Manipulation Behind the Run?
Send us a textMiguel Armaza stops by the New York Stock Exchange to interview Yoshi Yokokawa, CEO & Co-founder of Alpaca, an API-first brokerage infrastructure company that has raised over $100 million to serve more than 200 financial institutions across 40 countries, with 5 million brokerage accounts on their platform.In this episode, we discuss:Balancing data-driven decisions with entrepreneurial instinctWhy Alpaca's self-clearing model slashes dependencies, cuts costs, and accelerates innovationWhy API-first companies need a sandwich approach: bottom-up developer community and top-down enterprise salesWhy Japan's economic history shaped a generation of domestic entrepreneurs while global founders like Yoshi remain the exception… and lots more!Want more podcast episodes? Join me and follow Fintech Leaders today on Apple, Spotify, or your favorite podcast app for weekly conversations with today's global leaders that will dominate the 21st century in fintech, business, and beyond.Do you prefer a written summary? Check out the Fintech Leaders newsletter and join 75,000+ readers and listeners worldwide!Miguel Armaza is Co-Founder and General Partner of Gilgamesh Ventures, a seed-stage investment fund focused on fintech in the Americas. He also hosts and writes the Fintech Leaders podcast and newsletter.Miguel on LinkedIn: https://bit.ly/3nKha4ZMiguel on Twitter: https://bit.ly/2Jb5oBcFintech Leaders Newsletter: bit.ly/3jWIp
Welcome to Book of Lies Podcast! We have a drawing/contest to earn a free book of lies podcast mug for those listeners out in the universe, a winner will be annouced June 1st 2025. Email your entries to bookofliespodcast@gmail.comThis week we bring you an episode for American Greed streaming on Peacock. Joel Steiner was a convicted felon who was banned from selling securies in the 80's because he's a crook, that didnt stop him and his brothers from running and raising over 1 billions dollars from 30 thousand investors. Tap into this episode to watch out for a scam like this.We catch up on Sunni's birthday shindig, what we've been watching and more. Send us a message anytime, or just listen in to learn how to get a chance at this exclusive Book of Lies Podcast Mug.Subscribe and follow. Follow us on social @bookofliespodcast.
Tingly muffled crinkle while i ramble about alpacas and llamas differences and crinkle this dog toy ear to ear binaural real wholesome traditional ASMR triggers
All the Wool A Podcast for Hand Spinners, Knitters, and Yarn lovers
All the wool is a vlog all about handspinning yarn, processing wool, knitting, owning a wool mill, farm life and everything in between.To watch this episode on YouTubehttps://youtu.be/DO7my6WrTy4This episode starts in the mill at the picker as we blend up Shetland wool and Suri alpaca. Then more picking with Shetland wool and dyed up tussah silk. Off to the carder we go with these fibers and of course a little handspinning. There is also a quick visit to the farm to see the pigs, alpaca, June the pony and the sheep.Find me at:https://www.ewethfulfiberfarm.com/Blogging at http://www.beingewethful.com/Ravelry group: Ewethful Fiber Farm & MillJoin Ewethful's Patreon Communityhttps://www.patreon.com/EwethfulFiberMillTo ask me a questionhttps://docs.google.com/forms/d/e/1FAIpQLSdkoshX7grvAiOcNxwAlUqFskm-opVlE1h_L6jmdO-CvGX8kg/viewform?usp=sf_linkFree hand spinning resources - " Ewethful's Wool School"https://www.ewethfulfiberfarm.com/pages/wool-educationFor details and to purchase the online course to learn to handspin on wheelhttps://www.ewethfulfiberfarm.com/pages/lets-make-yarn-landing-pageFor details and to purchase the online course to learn to spin longdrawhttps://www.ewethfulfiberfarm.com/products/2256545Shop for Ewethful handspinning fibershttps://www.ewethfulfiberfarm.com/collectionsMentioned in this episode:Breakwater Beach VestDesigner: Irina Anikeevaknitted with Ewetopia YarnNYTimes articles re: pronatalismhttps://www.nytimes.com/2025/04/21/us/politics/trump-birthrate-proposals.htmlForeign Affairs magazine November/December 2024 editionArticle referenced: The Age of Depopulation: Surviving a World Gone GrayAuthor: Nicholas EberstadtOther sources I followNathan Tankushttps://www.nytimes.com/2025/04/10/business/economy/tariffs-apparel-american-made.htmlChad Scotthttps://www.youtube.com/@UCO1ydt_TOAZfwEgJpgOx2jQ Keds Economisthttps://www.youtube.com/@UC-L8bPaJ_aJznCLLtUHxnLA The Weekly Show with Jon Stewarthttps://www.youtube.com/@UCQlJ7XpBtiMLKNSd4RAJmRQ Reuters, Al Jazeera, NYTimes, NPR, Foreign AffairsAffiliate link:Wooleryhttps://woolery.com/?aff=352
Welcome to the podcast for alpaca people!We don't really know what is going on inside the head of an alpaca. What are they thinking? What do they think of us? It doesn't stop us trying to speak alpaca, though. And if you can spend time around your alpacas, you stand a better chance of connecting. When you do, it is a bit magical.Spring continues to gather pace and there are more signs of its arrival - the swallows, the cuckoo, the woodpecker and butterflies. What are you noticing where you are?Reminder: From May, the podcast frequency will change, and new episodes will arrive fortnightly rather than weekly.Thanks for listening and I hope you enjoyed it.If you would like to be in touch, please contact me by email - steve@alpacatribe.com - or leave me a voicemail from your browser.Alpaca Tribe is hosted and produced by Steve Heatherington of Good Podcasting Works, which is part of The Waterside (Swansea) LtdThis podcast uses the following third-party services for analysis: OP3 - https://op3.dev/privacy
Earth AI is discovering untapped critical metal deposits at half the cost in a fraction of the time. They combine machine learning and new modular drilling technology to move from detecting a prospect to drilling in just three to six months. They recently announced $20M in Series B funding in an oversubscribed round led by Tamarack Global and Cantos Ventures. Participating investors include Overmatch, Alpaca, Sparkwave Capital, Y Combinator, and Scrum Ventures.Roman is a YC alum, ex-PhD candidate at the University of Sydney, and geologist with 10 years of industry and research experience in Europe, the Middle and Far East, and Australia.Here are 6 topics we covered in the podcast:1. The Problem The world faces a projected $10T demand for critical metals by 2050, yet new mineral discoveries have declined by 70% over the last decade. Current exploration methods are expensive, slow, and increasingly inefficient.2. AI-Powered Discovery Roman's team developed a machine learning model trained on 50 years of exploration data. The AI identifies geological proxies—subtle clues that indicate where valuable minerals may lie beneath the surface.3. Vertical Integration After realizing the industry wasn't ready to adopt their tech, Earth AI built its own exploration and drilling systems. This vertical integration slashed costs and sped up testing, allowing them to confirm mineral targets rapidly.4. Breakthrough Results Their AI led to the first-ever discovery of magnetic nickel on Australia's East Coast. They've since found lithium, silver, and lead deposits, securing land cheaply and unlocking high-value prospects.5. Founder Wisdom Roman emphasizes sustainable work habits and a strong internal compass. He credits storytelling as crucial to convincing investors and the broader industry of Earth AI's potential.--
Welcome to the podcast for alpaca people!Podcast updateThe frequency of the Alpaca Tribe podcast will be changing from May 2025. It will become every two weeks rather than weekly. The alternate weeks will be filled by a new podcast called View From The Valley, which will focus on the 'other' things, leaving Alpaca Tribe to focus more on the alpacas. I would love to know what you think.Valley updateLake emptied and is back to half-full - weather doing its thing. The geese are nesting and the fox is around. The other birds are noisy. Spring is here, but the grass has not quite arrived. Recent rain will help with the catch-up.Freds is coping but being careful on the steep slopes. Desire lines and the right light levels help him negotiate the hill with his poor eyesight.Try to take advantage of any good weather you have and spend some time with your alpacas.LinksClaire Whitehead of Camelid Veterinary Services has just published a new book - The Camelid Care Handbook for Alpaca & Llama Owners: Foundations of Care (Volume 1). Comprehensive and practical - it is worth checking out.Thanks for listening and I hope you enjoyed it.If you would like to be in touch, please contact me by email - steve@alpacatribe.com - or leave me a voicemail from your browser.Alpaca Tribe is hosted and produced by Steve Heatherington of Good Podcasting Works, which is part of The Waterside (Swansea) LtdThis podcast uses the following third-party services for analysis: OP3 - https://op3.dev/privacy
EdTech technology start-ups are being urged to apply to the second year of the successful EdTech Accelerator Programme run by The Learnovate Centre and Amazon Web Services (AWS). Ten places are available on the 8-week programme, which will provide start-ups with mentoring by leading experts in business and product development as well as significant AWS cloud computing credits. Those who are accepted onto the programme also have a chance to win $100,000 USD worth of AWS products and services at the end of the 8 weeks. The EdTech Accelerator Programme is a collaboration between Learnovate, a leading global future of work and learning research hub in Trinity College Dublin, and AWS. AWS is a Patron Member of Learnovate joining a global community of leaders in learning and technology driving innovation in learning and education. This is the second year of the EdTech Accelerator Programme, which is open to early-stage start-ups that are developing digital solutions to problems related to education and wish to accelerate growth and scale operations. The deadline for applications is Friday, 4 April, 2025. Each participant in the programme may be eligible for significant credits for AWS cloud computing services across a range of areas, including compute power, database storage, content delivery and other functionality. Over the course of the 8 weeks, they will also receive 7 full-day workshops led by experts in learning, business and technology at AWS offices at Charlemont Square and Burlington Road in Dublin. Successful applicants will also receive free access to certified AWS training. All participants will present their solutions at the final 'Demo Day' showcase where a winner - chosen by Learnovate, AWS and industry judges - will receive $100,000 USD of AWS credits. Last year, ALPACA - which developed a game-based tool that identifies potential reading issues in children earlier - was announced as the winner. This year's Accelerator Programme begins on 29 April 2025 and finishes on 17 June 2025. The deadline for applications is 4 April 2025. Applications can be made via the website, learnovatecentre.org/startups. The programme aims to provide business mentoring, technical guidance on building and scaling in the cloud, as well as regulatory, data privacy and cybersecurity considerations. Specialist expertise will also be available on topics such as pilot studies, trials, pathways to market, public procurement and investor readiness. Eligibility criteria for the EdTech Accelerator Programme include: Companies must have identified a learning problem or need Applicants must have a working prototype for their solution or require help developing one The proposed digital solution must be one that improves learner outcomes, enhances workflow, or reduces costs related to learning Companies must be registered and based in the Republic of Ireland. Learnovate Director Nessa McEniff said: "We are thrilled to continue our partnership with AWS in fostering innovation among startups. The accelerator programme has been instrumental in providing emerging companies with the resources and mentorship needed to scale their solutions effectively. We look forward to supporting a new cohort of startups in 2025 as they navigate the challenges and opportunities in the EdTech landscape." Learnovate Commercialisation Manager Tom Pollock said: "The global EdTech market continues to rapidly expand, and Irish start-ups are key players worldwide, providing learning solutions that are used in companies, schools, universities and at home. Last year's Accelerator Programme was a huge success with those accepted onto it enjoying the benefits of expert business and technology mentoring. We are delighted to continue to provide this start-up programme in partnership with AWS to increase the support Learnovate already provides to Irish companies building products that fuse learning and technology." Stephen Bonney of AWS said: "We are delighted to be launching the seco...
Forget therapy dogs or cuddly cats, giant alpacas have been lifting the spirits of local residents in care homes in Wicklow and Wexford recently!It seems alpaca therapy is effective for both young and old, and Joe Phelan from K2 Alpacas joins Seán to tell more.
Forget therapy dogs or cuddly cats, giant alpacas have been lifting the spirits of local residents in care homes in Wicklow and Wexford recently!It seems alpaca therapy is effective for both young and old, and Joe Phelan from K2 Alpacas joins Seán to tell more.
Programa 5x117, amb Josep Rull. En aquest programa ens ve de gust fer la biografia d'una alpaca molt bufona que es diu Geronimo. Era tan preciosa que va dividir Anglaterra: uns l'estimaven molt i altres... doncs encara l'estimaven m
Jesús lavaba los pies de sus discípulos con un propósito, este era para que ellos puedan servir a los demás y ser parte del cambio. Jesús nos llama a servir a otros. Mensaje online del 23 de marzo de 2025 desde la iglesia Caminodevida - Lima, Perú por el pastor Victor Alpaca.
Join Howard Lindzon, Phil Pearlman, Michael Parekh, and special guest Yoshi Yokokawa, CEO of Alpaca, as they break down the latest market chaos. From Tesla's freefall and China's electric vehicle dominance to AI's growing role in fintech, this episode explores the future of investing, trading, and globalization. Yoshi shares insights on Alpaca's mission to power the next wave of brokerage innovation, the impact of de-globalization, and why APIs are revolutionizing global finance.
Skinny dipping below 8000 feet; band room duel; broken honor code of surprise party; Don Corleone of the dot matrix; cat heart conversation piece; calcified bone man in living room.Unlock the BONUS SCENE(S) at improv4humans.com and gain access to every episode of i4h, all ad-free, as well as TONS of exclusive new podcasts delving deeper into improv, the history of comedy, music and sci-fi.See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.
Episode 159 February 27, 2025 On the Needles 1:27 ALL KNITTING LINKS GO TO RAVELRY UNLESS OTHERWISE NOTED. Please visit our Instagram page @craftcookreadrepeat for non-Rav photos and info Baby hat, doodle knit directory by Jamie Lomax, Lemonade Shop simple sock in Dunks– DONE!! Bankhead hat by Susie Gorlay, Blue Sky Fibers Woolstok in Midnight Sea–DONE!! Succulents 2025 Blanket CAL by Mallory Krall, Hue Loco DK in Terrarium Llama llama duck by Adrienne Fong, C W D: Handcrafted Products for the Mind, Body & Soul BFL Alpaca Nylon Sock in Sutro Baths Filoli Cowl by Ksenia Naidyon, AVFKW Floating in Current and Marine Layer (70% Alpaca, 20% Silk, 10% Cashmere goat), AVFKW Wild Bloom in Quartz and Rain Cloud (41% Alpaca, 35% Silk, 13% Merino, 10% Yak)-- DONE!! On the Easel 10:15 Half-way of the Secret 100-Day! Klaus Mäkelä & Yuja Wang | Ravel Piano Concerto for the Left Hand in D major On the Table 18:28 Julia Turshen signing with Nina LaCour Mustard ponzu chicken and broccoli from Ali Slagle substack Boyfriend salmon from Justine Cooks Smitten Kitchen Marsala Meatballs A diy potato & leek crostada with hasty pie crust On the Nightstand 29:15 We are now a Bookshop.org affiliate! You can visit our shop to find books we've talked about or click on the links below. The books are supplied by local independent bookstores and a percentage goes to us at no cost to you! The Secret Hours by Mick Herron (audio) Everyone on This Train is a Suspect by Benjamin Stevenson (audio) All Fours by Miranda July The Blue Hour by Paula Hawkins The Blue Hour by Paula Hawkins All the Beauty in the World by Patrick Bringley Life Cycle of the Common Octopus by Emma Knight
Tina says they have a lot to cover! First things first, a moment of silence for Joann's closing down - a U.S. based craft store retailer that has been in business for over 80 years. Don't worry, Meg's mom is buying her quilting fabric online now, so she's all good! Tina discusses the fake funny story how Michaels and Joann's were once a married couple, fought about what a craft store should be, and they got divorced and made their own stores. Jess reminisces about "cosplay" days to grab fabrics for different costumes. Drea mentions she is making out well with the yarn sales from the liquidation. Meg offers the idea of giving turtles to random people to help with Drea's excessive crochet turtle problem. Tina asks our listeners to send in questions they may have about our different projects we work on such as yarn dyeing, pattern designing, and podcasting to help us create mini bonus episodes to sprinkle into our regular episode releases. You can send these questions in through our website using on contact form found at the bottom of all pages.Next, Tina brings up questions and topic suggestions the cast has received from Discord and email submissions. They go through lots of different questions and ideas, and also pick some to save for a full episode later on!Find out more about the podcast as well as our cast Meg, Drea, Tina, and Jess at our website pardonmystash.com. Our blog has information regarding our current patterns and yarns being used for projects. Leave us a comment on your thoughts on our episode blog posts, or through our social media accounts!This episode was sponsored by Jimmy Beans Wool found at jimmybeanswool.com.
Duncan Sarkies' latest work, Star Gazers: A political thriller, set within New Zealand's Alpaca Breeders' Organisation.
Episode 158 February 13, 2025 On the Needles 1:56 ALL KNITTING LINKS GO TO RAVELRY UNLESS OTHERWISE NOTED. Please visit our Instagram page @craftcookreadrepeat for non-Rav photos and info Hoodola by Laura Nelkin, Lady Dye Yarns DK in Notorious, Yarn Love Amy March DK in Beneath the Waves– DONE!! Succulents 2025 Blanket CAL by Mallory Krall, Hue Loco DK in Echevaria– DONE!! January Colorwork Cuff Club by Summer Lee, Lemonade Shop Simple Sock in Dunks– DONE!! February Colorwork Cuff Club by Summer Lee, Fibernymph Dyeworks Bounce in Love Bug and others– DONE!! Baby hat, doodle knit directory by Jamie Lomax, Lemonade Shop simple sock in Dunks Llama llama duck by Adrienne Fong, C W D: Handcrafted Products for the Mind, Body & Soul BFL Alpaca Nylon Sock in Sutro Baths Filoli Cowl by Ksenia Naidyon, AVFKW Floating in Current and Marine Layer (70% Alpaca, 20% Silk, 10% Cashmere goat), AVFKW Wild Bloom in Quartz and Rain Cloud (41% Alpaca, 35% Silk, 13% Merino, 10% Yak) Cortney's knitting: Pressed Flower Pullover by Amy Christoffers in Neighborhood Fiber Studio DK Ramblewood and Suri Loft Mondawmin (which is burgundy and hot pink). (Wintery Knitting!) On the Easel 18:18 Secret 100 Day Project, flip through coming soon Finished one Winter sketchbook, a Pith sketchbook. “Blue Series” drawings Florals, plus a field trip to the new SF Flower Market On the Table 22:57 First recipe test for America's Test kitchen! Sesame-feta meatballs with burst tomatoes Shallot, egg, bread,FETA BRINE, cumin, lamb Meatballs, shallot, cherry tomato, feta cubes chicken thighs + fries - by Julia Turshen Yogurt marinade chicken thighs yogurt, mayo, lemon juice, garlic, za'atar ½ for marinade, ½ for drizzle Real Texas Nachos Recipe Valentine Care Packages “Winter” tacos with shrimp and slaw from A Dish for All Seasons Bionaturae Sourdough Fusilli On the Nightstand 37:31 We are now a Bookshop.org affiliate! You can visit our shop to find books we've talked about or click on the links below. The books are supplied by local independent bookstores and a percentage goes to us at no cost to you! Be Ready When the Luck Happens by Ina Garten (audio). Sword Crossed by Freya Marke A Study in Drowning by Ava Reid The Au Pair Affair by Tessa Bailey Onyx storm by Rebecca Yarros Full Speed to a Crash Landing by Beth Revis How to Steal a Galaxy by Beth Revis The Pairing by Casey McQuiston Flower Philosophy by Anna Potter A Sea of Unspoken Things by Adrienne Young Shred Sisters by Betsy Lerner The Madness of Crowds by Louise Penny A World of Curiosities by Louise Penny
This was a fine Thursday crossword by Aidan Deshong and Jacob Reed, most notable for its novel theme. In most Thursday crosswords, something unexpected happens in the grid -- maybe letters are doubled up, displaced, hidden in black squares, etc. But today it is the clues that are camouflaged: the answer is straightforward once you've decoded said clue. We have all the deets inside, so have a listen, and please tell others about us on social media.Show note imagery: It's ALPACA v. LLAMA, who will win?? ... Probably the textile industry
Welcome to the podcast for alpaca people!Sitting with the alpacas you can zone out but still be present. A delight to spend time with them and notice what is happening in the valley.Thanks for listening and I hope you enjoyed it.If you would like to be in touch, please contact me by email - steve@alpacatribe.com - or leave me a voicemail from your browser.Alpaca Tribe is hosted and produced by Steve Heatherington of Good Podcasting Works, which is part of The Waterside (Swansea) LtdThis podcast uses the following third-party services for analysis: OP3 - https://op3.dev/privacy
In this special in between season's episode I have pulled one from the Patreon. Here we dwell on all our missed steps in 2023. What missed step would be complete without special guest stars like, Michael Bolton, Mutt Lange, Brother's Grimm, Russell Crowe, Disney, Eric Clapton, John Wayne, Paul McCartney, Uri Gellar, Pokémon, Joey Tempest, Europe, Nikki Sixx, Motley Crue, Duffer Brothers, The Stranger Things, Venus, Benny Hinn, Carlos Santana, Santa Claus, Artificial Intelligence, Alexa, Norwegian Black Metal, Lovecraft, Black Sabbath, The Book of Revelation, Dragons, Phoenician bath house, Nightrider, Shakespeare, first folio, Melania Trump, Quarto, 2015 Toyota Corolla, witchcraft, Knitting, Casseroles, Alpaca farms, Witchfinder General Mathew Hopkins, Jesus, Cotton Mather, Salem Witch Trials, Roseanne Barr, Jesus Christ Superstar, Vatican, Gorge Pell, Boys' choir, Genshin Impact, GUIZHONG, HEAVENLY PRINCIPLE, Sam Smith, #666 #SketchComedy #Sketch #Comedy #Sketch Comedy #Atheist #Science #History #Atheism #Antitheist #ConspiracyTheory #Conspiracy #Conspiracies #Sceptical #Scepticism #Mythology #Religion #Devil #Satan #Satanism #Satanist #Skeptic #Debunk #Illuminati #SatanIsMySuperhero #Podcast #funny #sketch #skit #comedy #comedyshow #comedyskits #HeavyMetal #weird #RomanEmpire #Rome #AncientRome #Romans #RomanEmperor #Animation #Anime
For this week's Atlantic Tales, Pat Flynn travelled to West Clare to meet the Italian couple who run their own Alpaca Farm. The Flying Alpaca was established by Barbara and Mike after they left their fulltime jobs and decided to settle in Ireland.
The Miskatonic Playhouse presents 'Alpaca in the Coffee' by John Hedge How would you react when the last person who'd ever use violence, turns violent? Something is very wrong with the coffee. Can the investigators unravel the mystery and confront a burgeoning sentience before it causes the city to tear itself apart? Alpaca in the coffee is a modern scenario set in Leith, part of Edinburgh in Scotland. It is an investigative mystery involving wyrd science, AI, and difficult choices. It is the third installment in the Alpaca Series. ---------- Keeper of Arcane Lore: John Hedge Vada Chakrabati: Newman Harper Greenman: Rina Haenze Edward Murphy: Pete Karla Nogacz: Stu --------- Scenario available here: https://www.drivethrurpg.com/en/product/465614/alpaca-in-the-coffee?affiliate_id=3500905 --------- Find us at www.MiskatonicPlayhouse.com Support us at ko-fi.com/MiskatonicPlayhouse
Welcome to the podcast for alpaca people!Happy New Year to you all. I have found my happy place to be with my alpacas. I hope it is yours too and you get a chance to spend some time with them.What will 2025 hold? More of the same? Different? How can we do better?No resolutions but a question to carry through the year. What will yours be?Peace and presence, SteveThanks for listening and I hope you enjoyed it.If you would like to be in touch, please contact me by email - steve@alpacatribe.com - or leave me a voicemail from your browser.Alpaca Tribe is hosted and produced by Steve Heatherington of Good Podcasting Works, which is part of The Waterside (Swansea) LtdThis podcast uses the following third-party services for analysis: OP3 - https://op3.dev/privacy
For the six weeks leading up to Christmas, Jodi Walker has a meet-cute with your Ringer favorites to binge the best original holiday movies that Hallmark, Lifetime, and Netflix have to offer. On the final day of Bingemas (also known as Christmas Day), Jodi and Rob Mahoney bring to you two movies filled with animal-based holiday love stories: ‘Happy Howlidays,' starring the dead-eyed winner of Hallmark's reality show to find a new holiday hunk ‘Finding Mr. Christmas' (6:35), and ‘Christmas on the Alpaca Farm,' featuring a man who would rather his entire life fall to ruin than make sweaters with blended fabrics (55:50). Thank you so much for binging with us this year! Host: Jodi Walker Guest: Rob Mahoney Producer: Sasha Ashall Learn more about your ad choices. Visit podcastchoices.com/adchoices
Dive into one of the most exciting and lucrative niches for travel advisors - destination weddings and group travel! Ashley Morris, founder of Alpaca Your Bags Travel, shares her journey of becoming a top destination wedding planner. From setting client expectations and handling budgets to marketing yourself as an expert, Ashley covers it all. Learn the benefits of having a repeatable workflow, partnering with the right vendors, and establishing boundaries to create a sustainable business. Ashley also discusses her role with Destination Wedding University, where she provides educational resources for aspiring destination wedding advisors. If you're looking to make destination weddings a profitable part of your business, this episode is full of strategies and tips to help you get there! About Ashley Morris: Ashley is the owner of award-winning Alpaca Your Bags Travel and is currently based in Kansas City, MO. She came to the travel industry after spending 15 years in media and advertising, working with clients like Target, Kohl's, Verizon and Spotify. She started Alpaca after a less-than-ideal experience planning her own destination wedding. After being frustrated with the planning process, she went to travel agent school and started her agency in 2015. Ashley built Alpaca into a million-dollar agency in under two years, all while still working at her advertising job AND commuting from KC to NYC twice a month. This is a woman who knows something about time management and working smarter, not harder! After having her own destination wedding, Ashley wanted to bring that joy and fun to others. She is a graduate of Will's first ever Destination Wedding University class, way back in 2018. Ashley and her husband, Matt, are fur parents to 2 dogs (chihuahua Ellie and beagle Jack) and 2 cats (Mork & Mindy). They are avid explorers and try to explore the world as much as possible, whether it's an African safari or towing their Airstream camper. alpacayourbags.com jointheplaybook.com Today we will cover: (04:30) Ashley's background and journey into travel (09:10) Destination Wedding University (12:15) Selecting a specialization; creating efficiency and profitability (16:55) The role of a destination wedding travel advisor (25:35) Can new advisors specialize in destination weddings? (40:15) Managing guest expectations and challenges (43:45) Handling price matching and contracts (46:15) Building strong tour operator partnerships (49:05) Establishing boundaries; using automation for repetitive tasks (58:45) Ashley's marketing efforts; creating a referral program JOIN THE NICHE COMMUNITY Enrollment opens on January 6th, 2025! SEMI CUSTOM BRAND KITS Browse our collection of pre-designed brand kits that are customized to your brand's unique personality by our talented Creative Director! _____ FOLLOW ALONG ON INSTAGRAM: instagram.com/tiquehq CHECK OUT OUR SERVICES & PROGRAMS: tiquehq.com
In episode #29 of Fintech Underground, our CEO, Yoshi, speaks with Mark Chahwan, Co-Founder and CEO of Sarwa about the following topics: 0:00 - Why Sarwa built Sarwa Trade as a way to give their customers more control over their own investing and trading decisions. 4:00 - Why Sarwa has focused on building a robust product suite before expanding into new markets. 10:00 - The impact of adding options to Sarwa's product suite and how options trading has been a growing trend. 15:00 - How COVID sparked a global interest in investing and asset ownership. 19:00 - The growing trend of increasing access to markets beyond the U.S. exchange. 29:00 - What Mark feels is the hidden gem of the fintech world. Thank you for listening to Alpaca's latest episode of Fintech Underground. You can find the YouTube video and transcript here: https://alpaca.markets/learn/how-sarwas-options-trading-and-product-diversity-has-helped-them-scale-quickly
It can happen to the best of us — classroom management deteriorating over time. Don't despair! By figuring out where the problems are, you can turn things back around. My guest is Claire English, who runs an incredible platform called The Unteachables, where she shows teachers how to manage their classrooms with confidence and calm. We talk about three reasons why classroom management can fall apart midway through a school year and what teachers can do to get things heading back in the right direction. Thanks to Scholastic Magazines+ and Alpaca for sponsoring this episode. For a full transcript of this episode, visit cultofpedagogy.com/pod and choose episode 240. Learn more from Claire through her courses*, That'll Teach 'Em and The Low Level Behaviour Bootcamp. and her membership site, The Behaviour Club,. *I am an affiliate of the Unteachables Academy. This means Cult of Pedagogy receives a commission on every purchase made through these links.
We do apologise cause you might actually laugh at some of these bad, bad jokes.
On this week's episode, we welcome comedian and podcaster, Carolina Hildago (No Dogs In Space), to chat about fan weddings, not driving in Los Angeles, the adorable Axolotl, and more!Be sure to get our new 'Ack Tuah' shirt in the Max Fun store.Or, grab an 'Ack Tuah' mug!Follow the podcast on Instagram and send us your dank memes!Check out Jesse's thrifted clothing store, Put This On.Go see Free With Ads and Judge John Hodgman LIVE at SF Sketchfest!Come see Judge John Hodgman: Road Court live in a town near you! Jesse and John will be all over the country so don't miss your change to see them. Check the events page to find out where! Follow brand new producer, Steven Ray Morris, on Instagram.Listen to See Jurassic Right!
Multitasking isn't great for our brains, it compromises our mental health, and ultimately it doesn't even work, but that doesn't stop many of us from trying to do it all the time. What we may not realize is that it also sneaks into our classrooms and interferes with learning, and it's happening in ways you may not even notice. In today's episode, I'm talking with cognitive scientist Megan Sumeracki about the pitfalls multitasking creates in schools, and what we can do to avoid them. -------------------------------------------- Thanks to The Gilder Lehrman Institute and Alpaca for sponsoring this episode. For a full transcript of this episode, visit cultofpedagogy.com/pod and choose episode 239.
Welcome to the podcast for alpaca people!Feeding alpacas can be fun and sometimes challenging. Always a great opportunity to catch up with how your herd is doing and what is going on. Also, it can add to the positive experiences they have of you as a safe person to be around.Of course, sometimes there is more than you bargain for.I have been musing about what makes a good alpaca. Particularly relevant if you are thinking of adding alpacas to your life or your herd. One factor among many is how sociable they are. Will they get on with the others you will be introducing them to? Are they going to be easy to manage? Often you can't be sure until later and see how they grow. We gather the information we can. Ask the relevant questions and make the best decision we can.What would you look for in an alpaca? Let me know by dropping me a line.Thanks for listening and I hope you enjoyed it.If you would like to be in touch, please contact me by email - steve@alpacatribe.com - or leave me a voicemail from your browser.Alpaca Tribe is hosted and produced by Steve Heatherington of Good Podcasting Works, which is part of The Waterside (Swansea) LtdThis podcast uses the following third-party services for analysis: OP3 - https://op3.dev/privacy
Apologies for lower audio quality; we lost recordings and had to use backup tracks. Our guests today are Anastasios Angelopoulos and Wei-Lin Chiang, leads of Chatbot Arena, fka LMSYS, the crowdsourced AI evaluation platform developed by the LMSys student club at Berkeley, which became the de facto standard for comparing language models. Arena ELO is often more cited than MMLU scores to many folks, and they have attracted >1,000,000 people to cast votes since its launch, leading top model trainers to cite them over their own formal academic benchmarks:The Limits of Static BenchmarksWe've done two benchmarks episodes: Benchmarks 101 and Benchmarks 201. One issue we've always brought up with static benchmarks is that 1) many are getting saturated, with models scoring almost perfectly on them 2) they often don't reflect production use cases, making it hard for developers and users to use them as guidance. The fundamental challenge in AI evaluation isn't technical - it's philosophical. How do you measure something that increasingly resembles human intelligence? Rather than trying to define intelligence upfront, Arena let users interact naturally with models and collect comparative feedback. It's messy and subjective, but that's precisely the point - it captures the full spectrum of what people actually care about when using AI.The Pareto Frontier of Cost vs IntelligenceBecause the Elo scores are remarkably stable over time, we can put all the chat models on a map against their respective cost to gain a view of at least 3 orders of magnitude of model sizes/costs and observe the remarkable shift in intelligence per dollar over the past year:This frontier stood remarkably firm through the recent releases of o1-preview and price cuts of Gemini 1.5:The Statistics of SubjectivityIn our Benchmarks 201 episode, Clémentine Fourrier from HuggingFace thought this design choice was one of shortcomings of arenas: they aren't reproducible. You don't know who ranked what and what exactly the outcome was at the time of ranking. That same person might rank the same pair of outputs differently on a different day, or might ask harder questions to better models compared to smaller ones, making it imbalanced. Another argument that people have brought up is confirmation bias. We know humans prefer longer responses and are swayed by formatting - Rob Mulla from Dreadnode had found some interesting data on this in May:The approach LMArena is taking is to use logistic regression to decompose human preferences into constituent factors. As Anastasios explains: "We can say what components of style contribute to human preference and how they contribute." By adding these style components as parameters, they can mathematically "suck out" their influence and isolate the core model capabilities.This extends beyond just style - they can control for any measurable factor: "What if I want to look at the cost adjusted performance? Parameter count? We can ex post facto measure that." This is one of the most interesting things about Arena: You have a data generation engine which you can clean and turn into leaderboards later. If you wanted to create a leaderboard for poetry writing, you could get existing data from Arena, normalize it by identifying these style components. Whether or not it's possible to really understand WHAT bias the voters have, that's a different question.Private EvalsOne of the most delicate challenges LMSYS faces is maintaining trust while collaborating with AI labs. The concern is that labs could game the system by testing multiple variants privately and only releasing the best performer. This was brought up when 4o-mini released and it ranked as the second best model on the leaderboard:But this fear misunderstands how Arena works. Unlike static benchmarks where selection bias is a major issue, Arena's live nature means any initial bias gets washed out by ongoing evaluation. As Anastasios explains: "In the long run, there's way more fresh data than there is data that was used to compare these five models." The other big question is WHAT model is actually being tested; as people often talk about on X / Discord, the same endpoint will randomly feel “nerfed” like it happened for “Claude European summer” and corresponding conspiracy theories:It's hard to keep track of these performance changes in Arena as these changes (if real…?) are not observable.The Future of EvaluationThe team's latest work on RouteLLM points to an interesting future where evaluation becomes more granular and task-specific. But they maintain that even simple routing strategies can be powerful - like directing complex queries to larger models while handling simple tasks with smaller ones.Arena is now going to expand beyond text into multimodal evaluation and specialized domains like code execution and red teaming. But their core insight remains: the best way to evaluate intelligence isn't to simplify it into metrics, but to embrace its complexity and find rigorous ways to analyze it. To go after this vision, they are spinning out Arena from LMSys, which will stay as an academia-driven group at Berkeley.Full Video PodcastChapters* 00:00:00 - Introductions* 00:01:16 - Origin and development of Chatbot Arena* 00:05:41 - Static benchmarks vs. Arenas* 00:09:03 - Community building* 00:13:32 - Biases in human preference evaluation* 00:18:27 - Style Control and Model Categories* 00:26:06 - Impact of o1* 00:29:15 - Collaborating with AI labs* 00:34:51 - RouteLLM and router models* 00:38:09 - Future of LMSys / ArenaShow Notes* Anastasios Angelopoulos* Anastasios' NeurIPS Paper Conformal Risk Control* Wei-Lin Chiang* Chatbot Arena* LMSys* MTBench* ShareGPT dataset* Stanford's Alpaca project* LLMRouter* E2B* DreadnodeTranscriptAlessio [00:00:00]: Hey everyone, welcome to the Latent Space podcast. This is Alessio, Partner and CTO in Residence at Decibel Partners, and I'm joined by my co-host Swyx, founder of Smol.ai.Swyx [00:00:14]: Hey, and today we're very happy and excited to welcome Anastasios and Wei Lin from LMSys. Welcome guys.Wei Lin [00:00:21]: Hey, how's it going? Nice to see you.Anastasios [00:00:23]: Thanks for having us.Swyx [00:00:24]: Anastasios, I actually saw you, I think at last year's NeurIPS. You were presenting a paper, which I don't really super understand, but it was some theory paper about how your method was very dominating over other sort of search methods. I don't remember what it was, but I remember that you were a very confident speaker.Anastasios [00:00:40]: Oh, I totally remember you. Didn't ever connect that, but yes, that's definitely true. Yeah. Nice to see you again.Swyx [00:00:46]: Yeah. I was frantically looking for the name of your paper and I couldn't find it. Basically I had to cut it because I didn't understand it.Anastasios [00:00:51]: Is this conformal PID control or was this the online control?Wei Lin [00:00:55]: Blast from the past, man.Swyx [00:00:57]: Blast from the past. It's always interesting how NeurIPS and all these academic conferences are sort of six months behind what people are actually doing, but conformal risk control, I would recommend people check it out. I have the recording. I just never published it just because I was like, I don't understand this enough to explain it.Anastasios [00:01:14]: People won't be interested.Wei Lin [00:01:15]: It's all good.Swyx [00:01:16]: But ELO scores, ELO scores are very easy to understand. You guys are responsible for the biggest revolution in language model benchmarking in the last few years. Maybe you guys want to introduce yourselves and maybe tell a little bit of the brief history of LMSysWei Lin [00:01:32]: Hey, I'm Wei Lin. I'm a fifth year PhD student at UC Berkeley, working on Chatbot Arena these days, doing crowdsourcing AI benchmarking.Anastasios [00:01:43]: I'm Anastasios. I'm a sixth year PhD student here at Berkeley. I did most of my PhD on like theoretical statistics and sort of foundations of model evaluation and testing. And now I'm working 150% on this Chatbot Arena stuff. It's great.Alessio [00:02:00]: And what was the origin of it? How did you come up with the idea? How did you get people to buy in? And then maybe what were one or two of the pivotal moments early on that kind of made it the standard for these things?Wei Lin [00:02:12]: Yeah, yeah. Chatbot Arena project was started last year in April, May, around that. Before that, we were basically experimenting in a lab how to fine tune a chatbot open source based on the Llama 1 model that I released. At that time, Lama 1 was like a base model and people didn't really know how to fine tune it. So we were doing some explorations. We were inspired by Stanford's Alpaca project. So we basically, yeah, grow a data set from the internet, which is called ShareGPT data set, which is like a dialogue data set between user and chat GPT conversation. It turns out to be like pretty high quality data, dialogue data. So we fine tune on it and then we train it and release the model called V2. And people were very excited about it because it kind of like demonstrate open way model can reach this conversation capability similar to chat GPT. And then we basically release the model with and also build a demo website for the model. People were very excited about it. But during the development, the biggest challenge to us at the time was like, how do we even evaluate it? How do we even argue this model we trained is better than others? And then what's the gap between this open source model that other proprietary offering? At that time, it was like GPT-4 was just announced and it's like Cloud One. What's the difference between them? And then after that, like every week, there's a new model being fine tuned, released. So even until still now, right? And then we have that demo website for V2 now. And then we thought like, okay, maybe we can add a few more of the model as well, like API model as well. And then we quickly realized that people need a tool to compare between different models. So we have like a side by side UI implemented on the website to that people choose, you know, compare. And we quickly realized that maybe we can do something like, like a battle on top of ECLMs, like just anonymize it, anonymize the identity, and that people vote which one is better. So the community decides which one is better, not us, not us arguing, you know, our model is better or what. And that turns out to be like, people are very excited about this idea. And then we tweet, we launch, and that's, yeah, that's April, May. And then it was like first two, three weeks, like just a few hundred thousand views tweet on our launch tweets. And then we have regularly double update weekly, beginning at a time, adding new model GPT-4 as well. So it was like, that was the, you know, the initial.Anastasios [00:04:58]: Another pivotal moment, just to jump in, would be private models, like the GPT, I'm a little,Wei Lin [00:05:04]: I'm a little chatty. That was this year. That was this year.Anastasios [00:05:07]: Huge.Wei Lin [00:05:08]: That was also huge.Alessio [00:05:09]: In the beginning, I saw the initial release was May 3rd of the beta board. On April 6, we did a benchmarks 101 episode for a podcast, just kind of talking about, you know, how so much of the data is like in the pre-training corpus and blah, blah, blah. And like the benchmarks are really not what we need to evaluate whether or not a model is good. Why did you not make a benchmark? Maybe at the time, you know, it was just like, Hey, let's just put together a whole bunch of data again, run a, make a score that seems much easier than coming out with a whole website where like users need to vote. Any thoughts behind that?Wei Lin [00:05:41]: I think it's more like fundamentally, we don't know how to automate this kind of benchmarks when it's more like, you know, conversational, multi-turn, and more open-ended task that may not come with a ground truth. So let's say if you ask a model to help you write an email for you for whatever purpose, there's no ground truth. How do you score them? Or write a story or a creative story or many other things like how we use ChatterBee these days. It's more open-ended. You know, we need human in the loop to give us feedback, which one is better. And I think nuance here is like, sometimes it's also hard for human to give the absolute rating. So that's why we have this kind of pairwise comparison, easier for people to choose which one is better. So from that, we use these pairwise comparison, those to calculate the leaderboard. Yeah. You can add more about this methodology.Anastasios [00:06:40]: Yeah. I think the point is that, and you guys probably also talked about this at some point, but static benchmarks are intrinsically, to some extent, unable to measure generative model performance. And the reason is because you cannot pre-annotate all the outputs of a generative model. You change the model, it's like the distribution of your data is changing. New labels to deal with that. New labels are great automated labeling, right? Which is why people are pursuing both. And yeah, static benchmarks, they allow you to zoom in to particular types of information like factuality, historical facts. We can build the best benchmark of historical facts, and we will then know that the model is great at historical facts. But ultimately, that's not the only axis, right? And we can build 50 of them, and we can evaluate 50 axes. But it's just so, the problem of generative model evaluation is just so expansive, and it's so subjective, that it's just maybe non-intrinsically impossible, but at least we don't see a way. We didn't see a way of encoding that into a fixed benchmark.Wei Lin [00:07:47]: But on the other hand, I think there's a challenge where this kind of online dynamic benchmark is more expensive than static benchmark, offline benchmark, where people still need it. Like when they build models, they need static benchmark to track where they are.Anastasios [00:08:03]: It's not like our benchmark is uniformly better than all other benchmarks, right? It just measures a different kind of performance that has proved to be useful.Swyx [00:08:14]: You guys also published MTBench as well, which is a static version, let's say, of Chatbot Arena, right? That people can actually use in their development of models.Wei Lin [00:08:25]: Right. I think one of the reasons we still do this static benchmark, we still wanted to explore, experiment whether we can automate this, because people, eventually, model developers need it to fast iterate their model. So that's why we explored LM as a judge, and ArenaHard, trying to filter, select high-quality data we collected from Chatbot Arena, the high-quality subset, and use that as a question and then automate the judge pipeline, so that people can quickly get high-quality signal, benchmark signals, using this online benchmark.Swyx [00:09:03]: As a community builder, I'm curious about just the initial early days. Obviously when you offer effectively free A-B testing inference for people, people will come and use your arena. What do you think were the key unlocks for you? Was it funding for this arena? Was it marketing? When people came in, do you see a noticeable skew in the data? Which obviously now you have enough data sets, you can separate things out, like coding and hard prompts, but in the early days, it was just all sorts of things.Anastasios [00:09:31]: Yeah, maybe one thing to establish at first is that our philosophy has always been to maximize organic use. I think that really does speak to your point, which is, yeah, why do people come? They came to use free LLM inference, right? And also, a lot of users just come to the website to use direct chat, because you can chat with the model for free. And then you could think about it like, hey, let's just be kind of like more on the selfish or conservative or protectionist side and say, no, we're only giving credits for people that battle or so on and so forth. Strategy wouldn't work, right? Because what we're trying to build is like a big funnel, a big funnel that can direct people. And some people are passionate and interested and they battle. And yes, the distribution of the people that do that is different. It's like, as you're pointing out, it's like, that's not as they're enthusiastic.Wei Lin [00:10:24]: They're early adopters of this technology.Anastasios [00:10:27]: Or they like games, you know, people like this. And we've run a couple of surveys that indicate this as well, of our user base.Wei Lin [00:10:36]: We do see a lot of developers come to the site asking polling questions, 20-30%. Yeah, 20-30%.Anastasios [00:10:42]: It's obviously not reflective of the general population, but it's reflective of some corner of the world of people that really care. And to some extent, maybe that's all right, because those are like the power users. And you know, we're not trying to claim that we represent the world, right? We represent the people that come and vote.Swyx [00:11:02]: Did you have to do anything marketing-wise? Was anything effective? Did you struggle at all? Was it success from day one?Wei Lin [00:11:09]: At some point, almost done. Okay. Because as you can imagine, this leaderboard depends on community engagement participation. If no one comes to vote tomorrow, then no leaderboard.Anastasios [00:11:23]: So we had some period of time when the number of users was just, after the initial launch, it went lower. Yeah. And, you know, at some point, it did not look promising. Actually, I joined the project a couple months in to do the statistical aspects, right? As you can imagine, that's how it kind of hooked into my previous work. At that time, it wasn't like, you know, it definitely wasn't clear that this was like going to be the eval or something. It was just like, oh, this is a cool project. Like Wayland seems awesome, you know, and that's it.Wei Lin [00:11:56]: Definitely. There's in the beginning, because people don't know us, people don't know what this is for. So we had a hard time. But I think we were lucky enough that we have some initial momentum. And as well as the competition between model providers just becoming, you know, became very intense. Intense. And then that makes the eval onto us, right? Because always number one is number one.Anastasios [00:12:23]: There's also an element of trust. Our main priority in everything we do is trust. We want to make sure we're doing everything like all the I's are dotted and the T's are crossed and nobody gets unfair treatment and people can see from our profiles and from our previous work and from whatever, you know, we're trustworthy people. We're not like trying to make a buck and we're not trying to become famous off of this or that. It's just, we're trying to provide a great public leaderboard community venture project.Wei Lin [00:12:51]: Yeah.Swyx [00:12:52]: Yes. I mean, you are kind of famous now, you know, that's fine. Just to dive in more into biases and, you know, some of this is like statistical control. The classic one for human preference evaluation is humans demonstrably prefer longer contexts or longer outputs, which is actually something that we don't necessarily want. You guys, I think maybe two months ago put out some length control studies. Apart from that, there are just other documented biases. Like, I'd just be interested in your review of what you've learned about biases and maybe a little bit about how you've controlled for them.Anastasios [00:13:32]: At a very high level, yeah. Humans are biased. Totally agree. Like in various ways. It's not clear whether that's good or bad, you know, we try not to make value judgments about these things. We just try to describe them as they are. And our approach is always as follows. We collect organic data and then we take that data and we mine it to get whatever insights we can get. And, you know, we have many millions of data points that we can now use to extract insights from. Now, one of those insights is to ask the question, what is the effect of style, right? You have a bunch of data, you have votes, people are voting either which way. We have all the conversations. We can say what components of style contribute to human preference and how do they contribute? Now, that's an important question. Why is that an important question? It's important because some people want to see which model would be better if the lengths of the responses were the same, were to be the same, right? People want to see the causal effect of the model's identity controlled for length or controlled for markdown, number of headers, bulleted lists, is the text bold? Some people don't, they just don't care about that. The idea is not to impose the judgment that this is not important, but rather to say ex post facto, can we analyze our data in a way that decouples all the different factors that go into human preference? Now, the way we do this is via statistical regression. That is to say the arena score that we show on our leaderboard is a particular type of linear model, right? It's a linear model that takes, it's a logistic regression that takes model identities and fits them against human preference, right? So it regresses human preference against model identity. What you get at the end of that logistic regression is a parameter vector of coefficients. And when the coefficient is large, it tells you that GPT 4.0 or whatever, very large coefficient, that means it's strong. And that's exactly what we report in the table. It's just the predictive effect of the model identity on the vote. The other thing that you can do is you can take that vector, let's say we have M models, that is an M dimensional vector of coefficients. What you can do is you say, hey, I also want to understand what the effect of length is. So I'll add another entry to that vector, which is trying to predict the vote, right? That tells me the difference in length between two model responses. So we have that for all of our data. We can compute it ex post facto. We added it into the regression and we look at that predictive effect. And then the idea, and this is formally true under certain conditions, not always verifiable ones, but the idea is that adding that extra coefficient to this vector will kind of suck out the predictive power of length and put it into that M plus first coefficient and quote, unquote, de-bias the rest so that the effect of length is not included. And that's what we do in style control. Now we don't just do it for M plus one. We have, you know, five, six different style components that have to do with markdown headers and bulleted lists and so on that we add here. Now, where is this going? You guys see the idea. It's a general methodology. If you have something that's sort of like a nuisance parameter, something that exists and provides predictive value, but you really don't want to estimate that. You want to remove its effect. In causal inference, these things are called like confounders often. What you can do is you can model the effect. You can put them into your model and try to adjust for them. So another one of those things might be cost. You know, what if I want to look at the cost adjusted performance of my model, which models are punching above their weight, parameter count, which models are punching above their weight in terms of parameter count, we can ex post facto measure that. We can do it without introducing anything that compromises the organic nature of theWei Lin [00:17:17]: data that we collect.Anastasios [00:17:18]: Hopefully that answers the question.Wei Lin [00:17:20]: It does.Swyx [00:17:21]: So I guess with a background in econometrics, this is super familiar.Anastasios [00:17:25]: You're probably better at this than me for sure.Swyx [00:17:27]: Well, I mean, so I used to be, you know, a quantitative trader and so, you know, controlling for multiple effects on stock price is effectively the job. So it's interesting. Obviously the problem is proving causation, which is hard, but you don't have to do that.Anastasios [00:17:45]: Yes. Yes, that's right. And causal inference is a hard problem and it goes beyond statistics, right? It's like you have to build the right causal model and so on and so forth. But we think that this is a good first step and we're sort of looking forward to learning from more people. You know, there's some good people at Berkeley that work on causal inference for the learning from them on like, what are the really most contemporary techniques that we can use in order to estimate true causal effects if possible.Swyx [00:18:10]: Maybe we could take a step through the other categories. So style control is a category. It is not a default. I have thought that when you wrote that blog post, actually, I thought it would be the new default because it seems like the most obvious thing to control for. But you also have other categories, you have coding, you have hard prompts. We consider that.Anastasios [00:18:27]: We're still actively considering it. It's just, you know, once you make that step, once you take that step, you're introducing your opinion and I'm not, you know, why should our opinion be the one? That's kind of a community choice. We could put it to a vote.Wei Lin [00:18:39]: We could pass.Anastasios [00:18:40]: Yeah, maybe do a poll. Maybe do a poll.Swyx [00:18:42]: I don't know. No opinion is an opinion.Wei Lin [00:18:44]: You know what I mean?Swyx [00:18:45]: Yeah.Wei Lin [00:18:46]: There's no neutral choice here.Swyx [00:18:47]: Yeah. You have all these others. You have instruction following too. What are your favorite categories that you like to talk about? Maybe you tell a little bit of the stories, tell a little bit of like the hard choices that you had to make.Wei Lin [00:18:57]: Yeah. Yeah. Yeah. I think the, uh, initially the reason why we want to add these new categories is essentially to answer some of the questions from our community, which is we won't have a single leaderboard for everything. So these models behave very differently in different domains. Let's say this model is trend for coding, this model trend for more technical questions and so on. On the other hand, to answer people's question about like, okay, what if all these low quality, you know, because we crowdsource data from the internet, there will be noise. So how do we de-noise? How do we filter out these low quality data effectively? So that was like, you know, some questions we want to answer. So basically we spent a few months, like really diving into these questions to understand how do we filter all these data because these are like medias of data points. And then if you want to re-label yourself, it's possible, but we need to kind of like to automate this kind of data classification pipeline for us to effectively categorize them to different categories, say coding, math, structure, and also harder problems. So that was like, the hope is when we slice the data into these meaningful categories to give people more like better signals, more direct signals, and that's also to clarify what we are actually measuring for, because I think that's the core part of the benchmark. That was the initial motivation. Does that make sense?Anastasios [00:20:27]: Yeah. Also, I'll just say, this does like get back to the point that the philosophy is to like mine organic, to take organic data and then mine it x plus factor.Alessio [00:20:35]: Is the data cage-free too, or just organic?Anastasios [00:20:39]: It's cage-free.Wei Lin [00:20:40]: No GMO. Yeah. And all of these efforts are like open source, like we open source all of the data cleaning pipeline, filtering pipeline. Yeah.Swyx [00:20:50]: I love the notebooks you guys publish. Actually really good just for learning statistics.Wei Lin [00:20:54]: Yeah. I'll share this insights with everyone.Alessio [00:20:59]: I agree on the initial premise of, Hey, writing an email, writing a story, there's like no ground truth. But I think as you move into like coding and like red teaming, some of these things, there's like kind of like skill levels. So I'm curious how you think about the distribution of skill of the users. Like maybe the top 1% of red teamers is just not participating in the arena. So how do you guys think about adjusting for it? And like feels like this where there's kind of like big differences between the average and the top. Yeah.Anastasios [00:21:29]: Red teaming, of course, red teaming is quite challenging. So, okay. Moving back. There's definitely like some tasks that are not as subjective that like pairwise human preference feedback is not the only signal that you would want to measure. And to some extent, maybe it's useful, but it may be more useful if you give people better tools. For example, it'd be great if we could execute code with an arena, be fantastic.Wei Lin [00:21:52]: We want to do it.Anastasios [00:21:53]: There's also this idea of constructing a user leaderboard. What does that mean? That means some users are better than others. And how do we measure that? How do we quantify that? Hard in chatbot arena, but where it is easier is in red teaming, because in red teaming, there's an explicit game. You're trying to break the model, you either win or you lose. So what you can do is you can say, Hey, what's really happening here is that the models and humans are playing a game against one another. And then you can use the same sort of Bradley Terry methodology with some, some extensions that we came up with in one of you can read one of our recent blog posts for, for the sort of theoretical extensions. You can attribute like strength back to individual players and jointly attribute strength to like the models that are in this jailbreaking game, along with the target tasks, like what types of jailbreaks you want.Wei Lin [00:22:44]: So yeah.Anastasios [00:22:45]: And I think that this is, this is a hugely important and interesting avenue that we want to continue researching. We have some initial ideas, but you know, all thoughts are welcome.Wei Lin [00:22:54]: Yeah.Alessio [00:22:55]: So first of all, on the code execution, the E2B guys, I'm sure they'll be happy to helpWei Lin [00:22:59]: you.Alessio [00:23:00]: I'll please set that up. They're big fans. We're investors in a company called Dreadnought, which we do a lot in AI red teaming. I think to me, the most interesting thing has been, how do you do sure? Like the model jailbreak is one side. We also had Nicola Scarlini from DeepMind on the podcast, and he was talking about, for example, like, you know, context stealing and like a weight stealing. So there's kind of like a lot more that goes around it. I'm curious just how you think about the model and then maybe like the broader system, even with Red Team Arena, you're just focused on like jailbreaking of the model, right? You're not doing kind of like any testing on the more system level thing of the model where like, maybe you can get the training data back, you're going to exfiltrate some of the layers and the weights and things like that.Wei Lin [00:23:43]: So right now, as you can see, the Red Team Arena is at a very early stage and we are still exploring what could be the potential new games we can introduce to the platform. So the idea is still the same, right? And we build a community driven project platform for people. They can have fun with this website, for sure. That's one thing, and then help everyone to test these models. So one of the aspects you mentioned is stealing secrets, stealing training sets. That could be one, you know, it could be designed as a game. Say, can you still use their credential, you know, we hide, maybe we can hide the credential into system prompts and so on. So there are like a few potential ideas we want to explore for sure. Do you want to add more?Anastasios [00:24:28]: I think that this is great. This idea is a great one. There's a lot of great ideas in the Red Teaming space. You know, I'm not personally like a Red Teamer. I don't like go around and Red Team models, but there are people that do that and they're awesome. They're super skilled. When I think about the Red Team arena, I think those are really the people that we're building it for. Like, we want to make them excited and happy, build tools that they like. And just like chatbot arena, we'll trust that this will end up being useful for the world. And all these people are, you know, I won't say all these people in this community are actually good hearted, right? They're not doing it because they want to like see the world burn. They're doing it because they like, think it's fun and cool. And yeah. Okay. Maybe they want to see, maybe they want a little bit.Wei Lin [00:25:13]: I don't know. Majority.Anastasios [00:25:15]: Yeah.Wei Lin [00:25:16]: You know what I'm saying.Anastasios [00:25:17]: So, you know, trying to figure out how to serve them best, I think, I don't know where that fits. I just, I'm not expressing. And give them credits, right?Wei Lin [00:25:24]: And give them credit.Anastasios [00:25:25]: Yeah. Yeah. So I'm not trying to express any particular value judgment here as to whether that's the right next step. It's just, that's sort of the way that I think we would think about it.Swyx [00:25:35]: Yeah. We also talked to Sander Schulhoff of the HackerPrompt competition, and he's pretty interested in Red Teaming at scale. Let's just call it that. You guys maybe want to talk with him.Wei Lin [00:25:45]: Oh, nice.Swyx [00:25:46]: We wanted to cover a little, a few topical things and then go into the other stuff that your group is doing. You know, you're not just running Chatbot Arena. We can also talk about the new website and your future plans, but I just wanted to briefly focus on O1. It is the hottest, latest model. Obviously, you guys already have it on the leaderboard. What is the impact of O1 on your evals?Wei Lin [00:26:06]: Made our interface slower.Anastasios [00:26:07]: It made it slower.Swyx [00:26:08]: Yeah.Wei Lin [00:26:10]: Because it needs like 30, 60 seconds, sometimes even more to, the latency is like higher. So that's one. Sure. But I think we observe very interesting things from this model as well. Like we observe like significant improvement in certain categories, like more technical or math. Yeah.Anastasios [00:26:32]: I think actually like one takeaway that was encouraging is that I think a lot of people before the O1 release were thinking, oh, like this benchmark is saturated. And why were they thinking that? They were thinking that because there was a bunch of models that were kind of at the same level. They were just kind of like incrementally competing and it sort of wasn't immediately obvious that any of them were any better. Nobody, including any individual person, it's hard to tell. But what O1 did is it was, it's clearly a better model for certain tasks. I mean, I used it for like proving some theorems and you know, there's some theorems that like only I know because I still do a little bit of theory. Right. So it's like, I can go in there and ask like, oh, how would you prove this exact thing? Which I can tell you has never been in the public domain. It'll do it. It's like, what?Wei Lin [00:27:19]: Okay.Anastasios [00:27:20]: So there's this model and it crushed the benchmark. You know, it's just like really like a big gap. And what that's telling us is that it's not saturated yet. It's still measuring some signal. That was encouraging. The point, the takeaway is that the benchmark is comparative. There's no absolute number. There's no maximum ELO. It's just like, if you're better than the rest, then you win. I think that was actually quite helpful to us.Swyx [00:27:46]: I think people were criticizing, I saw some of the academics criticizing it as not apples to apples. Right. Like, because it can take more time to reason, it's basically doing some search, doing some chain of thought that if you actually let the other models do that same thing, they might do better.Wei Lin [00:28:03]: Absolutely.Anastasios [00:28:04]: To be clear, none of the leaderboard currently is apples to apples because you have like Gemini Flash, you have, you know, all sorts of tiny models like Lama 8B, like 8B and 405B are not apples to apples.Wei Lin [00:28:19]: Totally agree. They have different latencies.Anastasios [00:28:21]: Different latencies.Wei Lin [00:28:22]: Control for latency. Yeah.Anastasios [00:28:24]: Latency control. That's another thing. We can do style control, but latency control. You know, things like this are important if you want to understand the trade-offs involved in using AI.Swyx [00:28:34]: O1 is a developing story. We still haven't seen the full model yet, but it's definitely a very exciting new paradigm. I think one community controversy I just wanted to give you guys space to address is the collaboration between you and the large model labs. People have been suspicious, let's just say, about how they choose to A-B test on you. I'll state the argument and let you respond, which is basically they run like five anonymous models and basically argmax their Elo on LMSYS or chatbot arena, and they release the best one. Right? What has been your end of the controversy? How have you decided to clarify your policy going forward?Wei Lin [00:29:15]: On a high level, I think our goal here is to build a fast eval for everyone, and including everyone in the community can see the data board and understand, compare the models. More importantly, I think we want to build the best eval also for model builders, like all these frontier labs building models. They're also internally facing a challenge, which is how do they eval the model? That's the reason why we want to partner with all the frontier lab people, and then to help them testing. That's one of the... We want to solve this technical challenge, which is eval. Yeah.Anastasios [00:29:54]: I mean, ideally, it benefits everyone, right?Wei Lin [00:29:56]: Yeah.Anastasios [00:29:57]: And people also are interested in seeing the leading edge of the models. People in the community seem to like that. Oh, there's a new model up. Is this strawberry? People are excited. People are interested. Yeah. And then there's this question that you bring up of, is it actually causing harm?Wei Lin [00:30:15]: Right?Anastasios [00:30:16]: Is it causing harm to the benchmark that we are allowing this private testing to happen? Maybe stepping back, why do you have that instinct? The reason why you and others in the community have that instinct is because when you look at something like a benchmark, like an image net, a static benchmark, what happens is that if I give you a million different models that are all slightly different, and I pick the best one, there's something called selection bias that plays in, which is that the performance of the winning model is overstated. This is also sometimes called the winner's curse. And that's because statistical fluctuations in the evaluation, they're driving which model gets selected as the top. So this selection bias can be a problem. Now there's a couple of things that make this benchmark slightly different. So first of all, the selection bias that you include when you're only testing five models is normally empirically small.Wei Lin [00:31:12]: And that's why we have these confidence intervals constructed.Anastasios [00:31:16]: That's right. Yeah. Our confidence intervals are actually not multiplicity adjusted. One thing that we could do immediately tomorrow in order to address this concern is if a model provider is testing five models and they want to release one, and we're constructing the models at level one minus alpha, we can just construct the intervals instead at level one minus alpha divided by five. That's called Bonferroni correction. What that'll tell you is that the final performance of the model, the interval that gets constructed, is actually formally correct. We don't do that right now, partially because we know from simulations that the amount of selection bias you incur with these five things is just not huge. It's not huge in comparison to the variability that you get from just regular human voters. So that's one thing. But then the second thing is the benchmark is live, right? So what ends up happening is it'll be a small magnitude, but even if you suffer from the winner's curse after testing these five models, what'll happen is that over time, because we're getting new data, it'll get adjusted down. So if there's any bias that gets introduced at that stage, in the long run, it actually doesn't matter. Because asymptotically, basically in the long run, there's way more fresh data than there is data that was used to compare these five models against these private models.Swyx [00:32:35]: The announcement effect is only just the first phase and it has a long tail.Anastasios [00:32:39]: Yeah, that's right. And it sort of like automatically corrects itself for this selection adjustment.Swyx [00:32:45]: Every month, I do a little chart of Ellim's ELO versus cost, just to track the price per dollar, the amount of like, how much money do I have to pay for one incremental point in ELO? And so I actually observe an interesting stability in most of the ELO numbers, except for some of them. For example, GPT-4-O August has fallen from 12.90
Books are one of the most powerful ways to learn about others and about ourselves. But for that learning to happen, we need a wide range of stories that represent a whole spectrum of people and lives. In many schools and classrooms, however, the offerings are far too narrow. Curating the kind of library that truly reflects the diversity of human experience takes time, intention, money, and good tools. This episode will help you make that happen. Joining me are three exceptional librarians — Cicely Lewis, Julia Torres, and Julie Stivers — who share their advice for building more inclusive collections. They also recommend a handful of outstanding titles to add to your shelves. Thanks to Scholastic Magazines+ and Alpaca for sponsoring this episode. For a full transcript of this episode, visit cultofpedagogy.com/pod and choose episode 237.
On this week's episode of The Rural Woman Podcast™, you'll meet Stéphanie Schiffgens, Marcia Cripps, Megan Harris, Abra Morawiec, Mickey Willenbring, Ashley Clark, Becca Matthews, Melissa Ballard, Colah B. Tawkin and Elaine Vandiver.Join us again for another special episode as we celebrate 5 years of The Rural Woman Podcast!! We are taking a look back and revisiting some of the stories that have been shared on The Rural Woman Podcast. In these throwback episodes, we will be highlighting these incredible, resilient women in agriculture. For full show notes, including links mentioned in the show, head over to wildrosefarmer.com/tbpt8. . .DISCUSSIONS THIS WEEK:[02:43] Diversified Farming in The Pacific Northwest with Stéphanie Schiffgens[07:00] In the Field of Agronomy with Marcia Cripps[09:39] Innovation and Creativity at The Farmers Market with Megan Harris[12:43] Pasture Raised Game Birds & Specialty Poultry with Abra Morawiec[16:17] Heritage Livestock & Indigenous Food Sovereignty with Combat Veteran, Mickey Willenbring[21:44] From Sap to Syrup: All About Maple Production with Ashley Clark[24:09] The Udder Milk: All about Dairy Sheep with Becca Matthews[26:30] Kentucky Grazed, Regeneratively Raised with Melissa Ballard from Bluegrass Beef[29:30] The Intersection of Black Culture and Horticulture with Colah B. Tawkin[33:10] Accidentally an Alpaca and Cut Flower Farm with Elaine Vandiver. . .This week's episode is brought to you by: Patreon . . .Looking to connect with fellow rural women and see Katelyn live? Don't miss out on the excitement! Click HERE to find an upcoming event near you!. . .Let's get SocialFollow The Rural Woman Podcast on Social MediaInstagram | FacebookSign up to get email updatesJoin our private Facebook group, The Rural Woman Podcast Community Connect with Katelyn on Instagram | Facebook | Twitter | Pinterest. . .Support the ShowPatreon | PayPal | Become a Show SponsorLeave a Review on Apple Podcasts | Take the Listener SurveyScreenshot this episode and share it on your socials!Tag...
Bye Bye K-Mart. Hero Cop. Morons in the News. Everyone Needs a Laugh. Too Good Looking. Small Plates. Talkback Callers. Can You Believe This S***? The Alpaca Haircut. From the Vault. Bob Shouts at the DMV.See omnystudio.com/listener for privacy information.
Dave Rubin is an American libertarian-conservative political commentator. He is the creator and host of The Rubin Report, a political talk show on YouTube and on the network BlazeTV. Visit Dave Rubin OFFICIAL Website here for more: https://daverubin.com/ Howie Mandel Does Stuff available on every podcast platform Howie Mandel Does Stuff Merchandise available on Amazon.com here https://www.amazon.com/shop/howiemandeldoesstuff Join the "Official Howie Mandel Does Stuff" Reddit: https://www.reddit.com/r/HowieMandelPodcast/ Thanks to our Sponsors: Whether it's Common Sense's travel size hand sanitizer with a fine mist spray that's TSA-friendly, personal size that foams and fits neatly in your car cup holder, or the bigger sizes for the heavy duty tasks of families or businesses, they all utilize Common Sense's unique bonded technology, leaving your hands feeling silky smooth for hours. Get 15% off with the code ‘HOWIE15' at smartcommonsense.com Warrior Alpaca Socks are made from premium Alpaca wool, which is perfect for sensitive skin. It's super soft, durable, and hypoallergenic. Best of all, Warrior Alpaca Socks are moisture-wicking, meaning they move moisture away from your feet to evaporate better, helping them stay dry and comfortable all day long. Get 20% off any pair you want with the code 'HOWIE' at warrioralpacasocks.com/howie GoWipe's Fruit & Veggie Wipes use natural bamboo fibers and a food- safe, non-toxic but super effective formula to keep surfaces clean. Just wipe your fruits,veggies, hands or anywhere fingers touch and boom – they're clean and ready to munch on. No water, no waiting – just wipe and bite! GoWipe is offering 10% off storewide with the code ‘HOWIE10' at gowipe.co Zest provides at-home blood tests, an app for tracking progress, and personalized longevity supplements based on your body's needs. Their app is so smart even my kids would be impressed. It's like having a personal health coach in your pocket, so know what's happening on the inside. You get 15% off storewide plus free shipping with code ‘HOWIE15' at zest.science Say Hello to our new house band Sunny and the Black Pack! Follow them here! YouTube: https://www.youtube.com/@BlackMediaPresents TikTok: https://www.tiktok.com/@blackmediapresents Spotify: https://open.spotify.com/artist/01uFmntCHwOW438t7enYOO?si=0Oc-_QJdQ0CrMkWii42BWA&nd=1&dlsi=a9792af062844b4f Facebook: https://www.facebook.com/SunnyAndTheBlackPack/ Instagram: https://www.instagram.com/blackmediapresents/ Twitch: https://www.twitch.tv/blackmediapresents Twitter: twitter.com/blackmedia Dave Rubin | Howie Mandel Does Stuff with Jackelyn Shultz #206 https://youtu.be/q1Mw91Xmddw @rubinreport @howiemandel @jackelynshultz