Podcasts about Keras

Neural network library

  • 383PODCASTS
  • 521EPISODES
  • 29mAVG DURATION
  • 1MONTHLY NEW EPISODE
  • Apr 3, 2025LATEST

POPULARITY

20172018201920202021202220232024


Best podcasts about Keras

Latest podcast episodes about Keras

The MAD Podcast with Matt Turck
Beyond Brute Force: Chollet & Knoop on ARC AGI 2, the Benchmark Breaking LLMs and the Search for True Machine Intelligence

The MAD Podcast with Matt Turck

Play Episode Listen Later Apr 3, 2025 60:45


In this fascinating episode, we dive deep into the race towards true AI intelligence, AGI benchmarks, test-time adaptation, and program synthesis with star AI researcher (and philosopher) Francois Chollet, creator of Keras and the ARC AGI benchmark, and Mike Knoop, co-founder of Zapier and now co-founder with Francois of both the ARC Prize and the research lab Ndea. With the launch of ARC Prize 2025 and ARC-AGI 2, they explain why existing LLMs fall short on true intelligence tests, how new models like O3 mark a step change in capabilities, and what it will really take to reach AGI.We cover everything from the technical evolution of ARC 1 to ARC 2, the shift toward test-time reasoning, and the role of program synthesis as a foundation for more general intelligence. The conversation also explores the philosophical underpinnings of intelligence, the structure of the ARC Prize, and the motivation behind launching Ndea — a ew AGI research lab that aims to build a "factory for rapid scientific advancement." Whether you're deep in the AI research trenches or just fascinated by where this is all headed, this episode offers clarity and inspiration.NdeaWebsite - https://ndea.comX/Twitter - https://x.com/ndeaARC PrizeWebsite - https://arcprize.orgX/Twitter - https://x.com/arcprizeFrançois CholletLinkedIn - https://www.linkedin.com/in/fcholletX/Twitter - https://x.com/fcholletMike KnoopX/Twitter - https://x.com/mikeknoopFIRSTMARKWebsite - https://firstmark.comX/Twitter - https://twitter.com/FirstMarkCapMatt Turck (Managing Director)LinkedIn - https://www.linkedin.com/in/turck/X/Twitter - https://twitter.com/mattturck(00:00) Intro (01:05) Introduction to ARC Prize 2025 and ARC-AGI 2 (02:07) What is ARC and how it differs from other AI benchmarks (02:54) Why current models struggle with fluid intelligence (03:52) Shift from static LLMs to test-time adaptation (04:19) What ARC measures vs. traditional benchmarks (07:52) Limitations of brute-force scaling in LLMs (13:31) Defining intelligence: adaptation and efficiency (16:19) How O3 achieved a massive leap in ARC performance (20:35) Speculation on O3's architecture and test-time search (22:48) Program synthesis: what it is and why it matters (28:28) Combining LLMs with search and synthesis techniques (34:57) The ARC Prize structure: efficiency track, private vs. public (42:03) Open source as a requirement for progress (44:59) What's new in ARC-AGI 2 and human benchmark testing (48:14) Capabilities ARC-AGI 2 is designed to test (49:21) When will ARC-AGI 2 be saturated? AGI timelines (52:25) Founding of NDEA and why now (54:19) Vision beyond AGI: a factory for scientific advancement (56:40) What NDEA is building and why it's different from LLM labs (58:32) Hiring and remote-first culture at NDEA (59:52) Closing thoughts and the future of AI research

Nick Kurniawan
MAKIN KERAS MAKIN LAKU KERAS (Mz11:5; Ams16:29; Yer51:46; Mal2:16)

Nick Kurniawan

Play Episode Listen Later Apr 1, 2025 9:23


Sabda Tuhan ingatkan bahwa makin sarat kekerasan, makin kabar berita jadi laku keras, meluas, membius dan membuas, hentikan propagandanya, berantas provokasinya.

makin keras sabda tuhan
Zonkuliah
ZK+AI - BM13 | 150512 | "Iman Tertinggi & Sifat Ahli Neraka" - Ustaz Shamsuri Ahmad

Zonkuliah

Play Episode Listen Later Feb 5, 2025 100:38


Kuliah Bahrul Mazi Jilid 13 yang berlangsung di Masjid Bandar Darul Aman, Jitra, Kedah pada 15 Mei 2012. ~ Berlaku satu zaman dahulu seorang lelaki beli tanah dari seseorang ~ Jumpa bekas ada emas dalam tanah ~ Aku beli tanah, bukan beli emas ~ Dalam dunia ni, walaupun ramai orang jahat, ada orang baik ~ Cadangan hakim untuk selesaikan kes ~ Satu masa dahulu ada orang yang sangat baik ~ Hadis yang cukup popular, hadis Maiz ~ Kisah sahabat yang sudah terlanjur berzina ~ Minta dihukum rejam ~ Allah tidak zalim untuk mengenakan hukuman di akhirat bagi yang sudah dikenakan hukum hudud di dunia ~ Islam bukan gelut nak hukum orang ~ Gugur hukum hudud jika ada syubhah ~ Undang-undang Islam bertujuan untuk menegakkan keadilan ~ "Hang makan ubat dak hari ni?" ~ Cara hukuman rejam kerana berzina dijalankan ~ Kes seorang perempuan yang minta dihukum kerana berzina ~ Adakah kita akan double standard jika ditakdirkan anak kita dikenakan hudud ~ Tahap keimanan para sahabat Nabi sangat tinggi ~ Sanggup sign mati -- BAHRUL MAZI JILID 13 MUKA SURAT 212 -- ~ Maksud takabbur terhadap Allah ~ Cakap yang menghina Allah ~ Hukum hudud tak sesuai diamalkan dalam negara ~ Siapa kita nak menolak hukum Allah ~ Contoh klasik Raja Namrud, yang nak berperang dengan tuhan ~ Firaun mengaku tuhan yang maha tinggi ~ Jangan hina orang yang taruh janggut ~ Menghina sunnah Nabi bermaksud menghina Nabi ~ Bukankah Nabi Muhammad pun manusia? Kenapa nak kena ikut dia? ~ Kaum Nabi Salleh yang samseng dengan Nabi ~ Muka berubah warna, datang jerkahan dari langit ~ Alangkah baiknya kalau ramadan kali ni, kita baca Quran dan faham makna ~ Maksud besar diri sesama manusia ~ Perasan diri pandai agama, orang lain bodoh ~ Ustaz tak mahu dengar ustaz lain mengajar ~ Sombong dan tinggi diri dah dikira dosa ~ Kita ni lemah, kecilnya kemampuan kita di atas dunia ~ Benci kepada perkara benar ~ Hang ingat semua orang yang bertudung tu bagus ka? ~ Bini tak takut kat aku ~ Kuat menderma, dan membantu orang tapi tak tutup aurat ~ Jangan kaitkan tudung dengan perbuatan salah individu ~ Khadam pakai tengkolok raja ~ Kebesaran adalah milik Allah ~ Pakai rida', selendang itu sunnah ~ Sifat yang menjadi hak Allah ~ Seksaan untuk orang yang takabbur ~ Suka ajuk orang yang gagap, akhirnya lekat kat kita ~ Orang yang cakap sombong, akhirnya akan jadi orang yang betul sombong ~ Pakaian mahal dan baik, boleh menimbulkan kesombongan ~ Jauhkan diri daripada sombong dengan harta benda mewah ~ Mudah sombong ini kena pada orang yang berpakaian elok ~ Debu daripada kaki kuda Nabi kena pada bapa munafik ~ Nabi sangat adil terhadap tindakannya ~ Nabi akhirnya termakan kepada kata-kata Abdullah Bin Ubai ~ Sedih sebab Nabi tak percaya ~ Seronoknya Quran dan Hadis, jadi pedoman hidup ~ Membesar diri adalah ahli neraka ~ Siapa yang akan menjadi ahli neraka (sifat ahli neraka) ~ Besar diri dan banyak daging ~ Keras tak boleh dibaca berbincang tak tentu punca ~ Bukan semua pendapat kita betul ~ Urusan orang Islam perlu diuruskan secara syura ~ Cara jalan pun nampak sombong ~ Orang yang sombong, Allah tak mahu bercakap dengan dia ~ Allah akan bercakap dengan ahli syurga ~ Penghuni syurga tak sangka dengan nikmat yang Allah bagi ~ Bayangkan seorang anak yang tak dapat redha mak ~ Nikmat paling besar yang ahli syurga akan dapat ~ Allah tak mahu tengok muka orang yang besar diri ~ Tiga orang yang Allah tidak pandang dengan rahmat -- Dapatkan External SSD 512GB Sempena 11 Tahun Zonkuliah : https://toyyibpay.com/Zonkuliah-External-SSD-512GB -- Sokong Projek Zonkuliah Dengan Menyumbang Ke Akaun Berikut : ➢ https://payment.tngdigital.com.my/sc/bDLnYClrWk ➢ MAYBANK (Produksi Zonkita) - 557250054584 ➢ PAYPAL - paypal.me/DanaZK --- ☑● Doakan Dimurahkan Rezeki dan Diberikan Kesihatan Yang Baik Untuk Kami Teruskan Projek ZonKuliah ☑● ✚ Untuk update terkini sila like Facebook Page kami : www.facebook.com/zonkuliah --- #Zonkuliah #UstazShamsuri #KuliahAgama

Narasipostmedia
Minuman Keras, Terus Beredar Meski Bahaya Tersebar

Narasipostmedia

Play Episode Listen Later Jan 19, 2025 7:19


Minuman Keras, Terus Beredar Meski Bahaya Tersebar Oleh. Ni'mah Fadeli(Kontributor NarasiPost.Com) Voice over talent: Dewi Nasjag NarasiPost.Com-Sebuah penelitian di Amerika mengungkapkan bahwa alkohol merupakan dalang di balik munculnya kanker. Temuan ini bukanlah hal yang mengejutkan, mengingat alkohol adalah salah satu barang haram dan pastinya tidak berguna bagi tubuh. Organisasi Kesehatan Dunia (WHO) pun telah mengingatkan bahwa tidak ada batas aman untuk mengonsumsi alkohol. Menurut WHO, bahkan sedikit saja alkohol masuk ke tubuh maka itu tetap dapat membahayakan kesehatan. (cnnindonesia.com, 04-01-2025). Naskah selengkapnya: ⁠https://narasipost.com/opini/01/2025/minuman-keras-terus-beredar-meski-bahaya-tersebar/ Terimakasih buat kalian yang sudah mendengarkan podcast ini, Follow us on: instagram: http://instagram.com/narasipost Facebook: https://www.facebook.com/narasi.post.9 Fanpage: Https://www.facebook.com/pg/narasipostmedia/posts/ Twitter: Http://twitter.com/narasipostx

amerika terus terimakasih meski bahaya keras organisasi kesehatan dunia who
CERITA PEMBELAJAR | Pengembangan Diri & Produktivitas
878 | Rahasia Storytelling Supaya Produkmu Laku Keras dan Laris Manis

CERITA PEMBELAJAR | Pengembangan Diri & Produktivitas

Play Episode Listen Later Jan 18, 2025 6:54


Ray Janson Radio
#469 CHEF!! KERJA KERAS SAMPAI BURNOUT?! WITH MAXIE MILLIAN | RAY JANSON RADIO

Ray Janson Radio

Play Episode Listen Later Dec 24, 2024 44:24


Maxie Millian kembali ke podcast! Pada episode kali ini, Maxie bercerita rencana dia memulai chapter baru di industry F&B. Maxie juga membahas pentingnya bekerja dengan happy namun tetap harus disiplin dan mencari peluang untuk memperoleh berbagai pengetahuan tanpa harus merusak kesehatan. Tonton video selengkapnya di #RayJansonRadio#469 CHEF!! KERJA KERAS SAMPAI BURNOUT?! WITH MAXIE MILLIAN | RAY JANSON RADIO Enjoy the show! Instagram: Maxie Millian: https://www.instagram.com/maxiemillian14/ DON'T FORGET TO LIKE AND SUBSCRIBE ! Ray Janson Radio is available on: Spotify: https://spoti.fi/2lEDF01 Apple Podcast: https://apple.co/2nhtizq Google Podcast: https://bit.ly/2laege8i Anchor App: https://anchor.fm/ray-janson-radio TikTok: https://www.tiktok.com/@rayjansonradio Let's talk some more: https://www.instagram.com/rayjanson #RayJansonRadio #FnBPodcast #Indonesia #SatuIndonesiaRasa

Radio Muhajir Project
Potongan dari Kajian Wanita 211. KERAS & KASAR = MERUSAK SEMUANYA

Radio Muhajir Project

Play Episode Listen Later Dec 24, 2024 1:05


Bismillah, 211. KERAS & KASAR = MERUSAK SEMUANYASimak Kajian Lengkapnya diYoutube Channel “Muhammad Nuzul Dzikri”Potongan dari Kajian Wanita

Radio Muhajir Project
Kajian Wanita 211. "KERAS & KASAR = MERUSAK SEMUANYA" | Kitab Al-Wabilush Shayyib

Radio Muhajir Project

Play Episode Listen Later Dec 24, 2024 49:55


Bismillah, 211. KERAS & KASAR = MERUSAK SEMUANYA Kajian Wanita Kitab Al-Wabilush Shayyib Pasal: Keutamaan shadaqah Halaman 95 Allah mencintai orang yang memililki sifat yang selaras dengan-Nya dengan segala keterbatasan makhluk tentunya, karena makhluk tidaklah sempurna dan Allah Maha Sempurna, dan jika sifat tersebut diperbolehkan. MATAN Dia-lah Allah Yang Maha Penyayang dan mencintai orang-orang yang penyayang, dan Dia hanya akan mengasihi hamba-hamba-Nya yang penyayang; Dia Maha Menutupi dan mencintai hamba-Nya yang menutupi aib atau kesalahan orang lain; Dia Maha Pemaaf dan mencintai hamba-Nya yang memaafkan orang lain; Dia Maha Pengampun dan mencintai orang yang diberikan ampunan; Dia Mahalembut dan mencintai sifat lembut dari hamba-Nya; Dia marah kepada orang yang kasar, keras, yang berjalan dengan sikap sombong; Dia Maha Pemurah dan mencintai sikap ramah; Dia Maha Murah hati dan mencintai sifat murah hati; Dia Mahabaik dan me-nyukai kebaikan dan para pelakunya; Dia Mahaadil dan mencintai keadilan; Maha Menerima udzur dari orang-orang yang memiliki udzur dan mencintai orang yang menerima udzur dari hamba-hamba-Nya; Dia memberikan balasan kepada hamba-hamba-Nya sesuai dengan sifat-sifat ini. Siapa yang memaafkan maka ia akan dimaafkan; siapa yang memberikan ampunan maka akan diampunkan; siapa yang memberikan izin maka akan diizinkan; siapa yang lemah lembut kepada hamba-Nya maka ia akan dilemahlembuti; siapa yang menyayangi makhluk-Nya maka ia akan disayangi; siapa yang berlaku baik kepada mereka maka akan diperlakukan secara baik; siapa yang derma kepada mereka maka akan diperlakukan dengan derma juga; siapa yang memberikan manfaat kepada mereka maka akan diberi manfaat baginya; siapa yang menutup aib mereka maka akan ditutup pula aibnya; siapa yang memaafkan mereka maka akan dimaafkan; siapa yang menyelidiki cacat mereka maka akan diselidiki juga cacatnya; siapa yang menghalangi kebaikan mereka maka kebaikannya akan dihalangi; siapa yang rindu kepada Allah maka Allah akan rindu kepadanya; siapa yang melakukan makar maka ia akan diberi makar; siapa yang menipu maka ia akan ditipu; siapa yang bermu'amalah kepada makhluk-Nya dengan satu sifat maka Allah akan bermu'amalah kepadanya dengan sifat tersebut di dunia dan di akhirat. Allah akan memperlakukan hamba-Nya sesuai dengan perlakuan hamba itu kepada makhluk-Nya.

Radio Muhajir Project
Potongan dari Kajian Wanita 210. "ALLAH MEMBENCI WANITA YANG KASAR & KERAS"

Radio Muhajir Project

Play Episode Listen Later Dec 16, 2024 0:58


Bismillah, 210. ALLAH MEMBENCI WANITA YANG KASAR & KERASSimak Kajian Lengkapnya diYoutube Channel “Muhammad Nuzul Dzikri”Potongan dari Kajian Wanita

Radio Muhajir Project
Kajian Wanita 210. "ALLAH MEMBENCI WANITA YANG KASAR & KERAS" | Kitab Al-Wabilush Shayyib

Radio Muhajir Project

Play Episode Listen Later Dec 16, 2024 23:39


Bismillah, 210. ALLAH MEMBENCI WANITA YANG KASAR & KERAS Kajian Wanita Kitab Al-Wabilush Shayyib Pasal: Keutamaan shadaqah Halaman 95 Allah mencintai orang yang memililki sifat yang selaras dengan-Nya dengan segala keterbatasan makhluk tentunya, karena makhluk tidaklah sempurna dan Allah Maha Sempurna, dan jika sifat tersebut diperbolehkan. MATAN Dia-lah Allah Yang Maha Penyayang dan mencintai orang-orang yang penyayang, dan Dia hanya akan mengasihi hamba-hamba-Nya yang penyayang; Dia Maha Menutupi dan mencintai hamba-Nya yang menutupi aib atau kesalahan orang lain; Dia Maha Pemaaf dan mencintai hamba-Nya yang memaafkan orang lain; Dia Maha Pengampun dan mencintai orang yang diberikan ampunan; Dia Mahalembut dan mencintai sifat lembut dari hamba-Nya; Dia marah kepada orang yang kasar, keras, yang berjalan dengan sikap sombong; Dia Maha Pemurah dan mencintai sikap ramah; Dia Maha Murah hati dan mencintai sifat murah hati; Dia Mahabaik dan me-nyukai kebaikan dan para pelakunya; Dia Mahaadil dan mencintai keadilan; Maha Menerima udzur dari orang-orang yang memiliki udzur dan mencintai orang yang menerima udzur dari hamba-hamba-Nya; Dia memberikan balasan kepada hamba-hamba-Nya sesuai dengan sifat-sifat ini. Siapa yang memaafkan maka ia akan dimaafkan; siapa yang memberikan ampunan maka akan diampunkan; siapa yang memberikan izin maka akan diizinkan; siapa yang lemah lembut kepada hamba-Nya maka ia akan dilemahlembuti; siapa yang menyayangi makhluk-Nya maka ia akan disayangi; siapa yang berlaku baik kepada mereka maka akan diperlakukan secara baik; siapa yang derma kepada mereka maka akan diperlakukan dengan derma juga; siapa yang memberikan manfaat kepada mereka maka akan diberi manfaat baginya; siapa yang menutup aib mereka maka akan ditutup pula aibnya; siapa yang memaafkan mereka maka akan dimaafkan; siapa yang menyelidiki cacat mereka maka akan diselidiki juga cacatnya; siapa yang menghalangi kebaikan mereka maka kebaikannya akan dihalangi; siapa yang rindu kepada Allah maka Allah akan rindu kepadanya; siapa yang melakukan makar maka ia akan diberi makar; siapa yang menipu maka ia akan ditipu; siapa yang bermu'amalah kepada makhluk-Nya dengan satu sifat maka Allah akan bermu'amalah kepadanya dengan sifat tersebut di dunia dan di akhirat. Allah akan memperlakukan hamba-Nya sesuai dengan perlakuan hamba itu kepada makhluk-Nya.

Parenting Podcast
"Ini Bedanya Anak Strong Willed dan Anak Keras Kepala" oleh Hanlie Muliani, M.Psi

Parenting Podcast

Play Episode Listen Later Dec 12, 2024 2:33


"Kenapa sih anak saya susah diatur, termasuk strong willed atau keras kepala? Kenali dulu bedanya dengan dengar audio ini, Moms & Dads. Kasih opini di komentar dong, kalau anak Anda punya ciri-ciri yang mana? "

Lentera Malam (Podcast Horor)
615 PERINGATAN KERAS!! JANGAN PERNAH MASUK KE TEMPAT INI SECARA ILEGAL

Lentera Malam (Podcast Horor)

Play Episode Listen Later Dec 12, 2024 89:15


Berniat liburan ke tempat wisata cagar alam, narasumber kali ini justru terjebak semalaman di goa jepang dengan kondisi temannya yang hampir patah tulang. Saat bermalam banyak sosok penampakan yang dilihatnya. Bagaimana kisah selengkapnya? Simak video berikut, jangan lupa berikan like dan komentarnya Copyright 2024, Lentera Malam --- Support this podcast: https://podcasters.spotify.com/pod/show/lentera-malam8/support

TestTalks | Automation Awesomeness | Helping YOU Succeed with Test Automation
AI Impact in Software Testing with Gaurav Mittal

TestTalks | Automation Awesomeness | Helping YOU Succeed with Test Automation

Play Episode Listen Later Nov 24, 2024 28:40


Today, Gaurav Mittal, an expert data science manager with over 18 years of experience, joins the podcast. In this episode, Gaurav shares his journey from manual to automation testing and delves into AI's revolutionary impact on software testing careers. Add visual checks toi your tests now: https://testguild.me/vizstack We'll explore how open-source AI models and libraries like TensorFlow and Keras make powerful tools accessible without the price tag and discuss the crucial role of retraining machine learning models to adapt to dynamic data. Gaurav will highlight the substantial benefits of automation in categorizing emails and its time-saving prowess. We'll also uncover the advantages of shift-left testing with AI, enhancing efficiency in the CI/CD pipeline and fostering collaboration among QA teams, developers, and project managers. Moreover, Gaurav offers a comparative insight between Selenium and the newer Playwright, advocating for the latter's superior performance. Throughout the episode, Gaurav emphasizes the practicality of AI as an assistant rather than a necessity in automation efforts. He also provides hands-on advice for integrating open-source AI models into your processes. Stay tuned for actionable tips and incredible insights on utilizing AI to elevate your automation game—all this and more, right here on the TestGuild Automation Podcast!  

VOA This Morning Podcast - Voice of America | Bahasa Indonesia
VOA This Morning "Negara G20 Sepakat Akan Pajaki Orang-orang Super Kaya; Kebijakan Baru Nuklir Rusia Tuai Reaksi Keras" - November 20, 2024

VOA This Morning Podcast - Voice of America | Bahasa Indonesia

Play Episode Listen Later Nov 19, 2024 15:48


Meski hasilkan komunike, para pemimpin G20 tak bisa sepakat soal sejumlah isu, termasuk Ukraina dan Gaza. Sementara itu, Presiden Putin tandatangani doktrin nuklir yang menyebut serangan konvensional negara yang didukung kekuatan nuklir terhadap Rusia akan dianggap serangan bersama terhadapnya.

SBS Indonesian - SBS Bahasa Indonesia
Finding a bank account that works hard for you - Menemukan rekening bank yang bekerja keras untuk Anda

SBS Indonesian - SBS Bahasa Indonesia

Play Episode Listen Later Nov 15, 2024 8:35


If you have a job, receive government benefits or want to pay your bills easily you'll need a bank account. You may even need more than one. To join the 20 million customers who hold Australian bank accounts, take some time to find one that best suits your needs. - Jika Anda memiliki pekerjaan, menerima tunjangan pemerintah atau ingin membayar tagihan Anda dengan mudah, Anda akan memerlukan rekening bank. Anda bahkan mungkin membutuhkan lebih dari satu. Untuk bergabung dengan 20 juta pelanggan yang memiliki rekening bank Australia, luangkan waktu untuk menemukan satu yang paling sesuai dengan kebutuhan Anda.

Benstage Podcast
Eps 20 - Ubah Mindset Kerja Keras Menjadi "In sya Allah Ada Rejekinya"

Benstage Podcast

Play Episode Listen Later Nov 2, 2024 4:59


Window of Opportunity - A Stargate Rewatch Podcast
Stargate Atlantis - Childhood's End

Window of Opportunity - A Stargate Rewatch Podcast

Play Episode Listen Later Oct 10, 2024 51:08


John and his team run into a society where no one is older than 24 in Childhood's End. We do have to wonder about what actually happens to a jumper if it loses power mid-flight. Wouldn't it just plummet straight to the ground since it doesn't have wings like a plane to help it glide? But it's a TV show so it all works out fine. It seems odd that the knowledge of the shield generator device has been lost to time even though the sacrifices continue with how connected the two things seem to be. And also, why isn't making sure the shield keeps working part of their culture if that's the thing that is keeping the people safe? We love how Weir puts Rodney in his place this week with what he's done to these kids and their protection from the wraith. Please tell me I'm not the only one who sees a little chemistry between John and Keras. There's something there right?!?! Rachael has a final thought on this episode!! INSTAGRAM: SG_Rewatch THREADS: SG_Rewatch DISCORD: https://discord.gg/65kMPzBuaN EMAIL: woosgrewatch@gmail.com

AWR Indonesian - Sabbath School Lesson

Hati para pemimpin dan pemuka agama yang jahat, kepercayaan yang salah bahwa Mesias membebaskan mereka dari Roma, kecintaan pada kekuasaan, kurangnya penyerahan diri kepada Allah, semua ini menyebabkan mereka menolak kebenaran di hadapan mereka.

AWR in Indonesian -Pelajaran Sekolah Sabat

Hati para pemimpin dan pemuka agama yang jahat, kepercayaan yang salah bahwa Mesias membebaskan mereka dari Roma, kecintaan pada kekuasaan, kurangnya penyerahan diri kepada Allah, semua ini menyebabkan mereka menolak kebenaran di hadapan mereka.

Podcast Jurnal Lembu
#411° KURANG KERAS HIDUP LOE: HIDUP BARENG UTANG

Podcast Jurnal Lembu

Play Episode Listen Later Sep 17, 2024 14:13


Selamat mendengarkan dan semoga bisa ambil hikmahnya guys!

Radio Rodja 756 AM
Azab yang Sangat Keras untuk Orang-Orang Kafir – Tafsir Surah Ali Imran 56-57

Radio Rodja 756 AM

Play Episode Listen Later Aug 29, 2024 55:37


Azab yang Sangat Keras untuk Orang-Orang Kafir – Tafsir Surah Ali Imran 56-57 adalah kajian tafsir Al-Quran yang disampaikan oleh Ustadz Abu Yahya Badrusalam, Lc. Kajian ini beliau sampaikan di Masjid Al-Barkah, komplek studio Radio Rodja dan Rodja TV pada Selasa, 22 Safar 1446 H / 27 Agustus 2024 M. Download kajian sebelumnya: Menghadapi Makar Musuh-Musuh […] Tulisan Azab yang Sangat Keras untuk Orang-Orang Kafir – Tafsir Surah Ali Imran 56-57 ditampilkan di Radio Rodja 756 AM.

Radio Muhajir Project
Riyaadhush Shaalihiin 1492. "MENDAKWAHI ORANG TUA YANG KERAS"

Radio Muhajir Project

Play Episode Listen Later Aug 13, 2024 40:41


Bismillah, 1492. MENDAKWAHI ORANG TUA YANG KERAS Riyaadhush Shaalihiin Bab 45 | Mengunjungi & mengundang orang-orang baik Tanya jawab === Saksikan Hasil RekamanKajian Rutin Serial Riyaadhush Shaalihiin bersama:

orang keras bismillah ustadz muhammad nuzul dzikri
Renungan Anak GKY Mabes
Jangan Keras Kepala (11 Agustus)

Renungan Anak GKY Mabes

Play Episode Listen Later Aug 10, 2024 3:05


Hai Wonder Kids, kembali dalam renungan anak GKY Mangga Besar. Judul renungan hari ini adalah   JANGAN KERAS KEPALA   Firman Tuhan dalam YAKOBUS 4:10 berkata -Rendahkanlah dirimu di hadapan Tuhan, dan Ia akan meninggikan kamu.   Wonder Kids, penting bagi kita untuk memanfaatkan waktu untuk saat teduh, yaitu memikirkan tentang TUHAN dan merenungkan Firman-Nya, serta bagaimana Firman Tuhan dapat mengubah hidup kita. Tapi jika kamu keras kepala dan hanya mau mengikuti keinginanmu sendiri, maka kamu akan sulit melakukan saat teduh. Kenapa? Karena TUHAN akan menggunakan saat teduhmu untuk mengungkapkan hal-hal yang perlu kamu ubah dari karaktermu. TUHAN akan tetap menunjukkan kepadamu sampai kamu menyetujui bahwa ini adalah masalah yang perlu dibereskan.   Jika kamu tidak bersedia memperbaiki apa yang TUHAN perlihatkan kepadamu, mungkin kamu tergoda untuk berhenti saat teduh supaya kamu tidak perlu mendengar teguran dari TUHAN. Tapi Wonder Kids, ini adalah kesalahan yang besar! Jangan biarkan dosa dan sikap keras kepala menjauhkanmu dari TUHAN. Bertobatlah, dan akuilah bahwa kamu salah dan TUHAN benar. Nikmatilah waktu bersama TUHAN yang begitu mengasihimu sehingga TUHAN akan memberitahumu bagaimana kamu bisa dibebaskan dari masalah yang membuatmu jauh dari TUHAN.   Mari kita berdoa,   TUHAN, mohon tunjukkan kepadaku dosa-dosa yang ada di dalam hidupku. Tolong aku untuk berpaling dari dosa-dosaku dan agar aku mendekat kepada-Mu, didalam nama Tuhan Yesus aku berdoa  Amin.     Wonder kids, HARI INI, INGAT UNTUK SAAT TEDUH DAN MEMIKIRKAN TENTANG TUHAN DAN FIRMAN-NYA. Tuhan Yesus memberkati.

Zonkuliah
BM13 | 210212 | "Dihalau Dari Al Kautsar & Jangan Bertengkar" - Ustaz Shamsuri Ahmad

Zonkuliah

Play Episode Listen Later Aug 5, 2024 90:39


Kuliah Bahrul Mazi Jilid 13 yang berlangsung di Masjid Bandar Darul Aman, Jitra, Kedah pada 21hb Februari 2012. ~ Antara orang yang awal mati syahid dalam perang Uhud ~ Sahabat yang kematian ayah, terpaksa tanggung hutang dan adik-adik perempuan ~ Nabi nampak sahabat sedih dan susah hati ~ Sahabat sensitif, nampak Nabi lapar ~ Tanggungan ramai dan tinggal hutang banyak ~ Berita baik tentang Allah, jumpa dengan bapa Jabir ~ Allah bercakap dengan hambaNya disebalik hijab ~ Manusia yang dipilih Allah untuk bercakap dengan Allah secara berdepan ~ Peluang untuk minta apa saja, dan Allah akan bagi ~ "Wahai tuhanku, bagi aku hidup lagi sekali untuk syahid kali kedua" ~ Manusia yang telah melalui kematian, tidak akan hidup kembali ke dunia ~ Mutawif Indonesia, bawa cerita merepek, roh balik ke rumah ~ Bercakap jangan sampai melanggar naf hadis dan Quran ~ Jangan berbohong atas nama Nabi ~ Jangan kamu sangka orang yang mati syahid itu mati ~ Mulianya orang yang mati dengan niat nak cari keredhaan Allah ~ Nabi solat jenazah kepada orang yang mati syahid selepas 8 tahun ~ Solat jenazah yang sangat syahdu, penuh kesedihan dan umpama perpisahan ~ Nabi berkhutbah dan bagitahu Nabi akan meninggal dah ~ Nabi akan menjadi saksi para sahabat ~ Janji telaga Nabi Al-Khausar ~ Ada orang yang dihalau malaikat ketika nak mendekati telaga Nabi -- BAHRUL MAZI JILID 13 MUKA SURAT 203 -- ~ Bab tak bagusnya lawan bertengkar ~ Atas sebab apa pun tak perlu bertengkar ~ Islam tak suka berbalah ~ Orang laki rimas dengan dengar isteri marah anak ~ Masalah nak tegakkan benang yang basah ~ Tau dia salah, tapi nak menang juga ~ Istana di pinggir syurga ~ Cara membangkitkan pertengkaran ~ Ada orang yang suka "complain" ~ Ustaz Shamsuri, baca hadis pun baris tak betul ~ Cari pasal untuk mencacatkan orang lain ~ Cakap macam lain, tapi dalam hati macam lain ~ Masalah tersasul ni semua orang hadapi ~ "Hang mengajar pun kami tidur.." ~ Manusia semua ada cacat cela ~ Ganjaran istana di tengah syurga ~ Cara untuk elakkan pertengkaran ~ "Saya ingat, ustaz dok pembohong orang ni.." ~ Cara dan teknik untuk buat teguran ~ Ganjaran untuk orang yang baik akhlak ~ Berkelahi itu tak bagus ~ Cukup seseorang itu dianggap berdosa, jika kaki berkelahi ~ Kaki gaduh dalam mesyuarat ~ Saja nak cari pasal untuk huru harakan majlis ~ Orang yang nak mencetuskan ketegangan tak perlu ramai dalam masyarakat ~ Keras kepala, dia sorang saja betul ~ Prinsip dalam hidup, mesti nak bangkang orang ~ Bercakap biar untuk cari kebenaran ~ Bagaimana nak perjuangkan hak daripada dizalimi ~ Lawan untuk tuntut hak dibenarkan, jangan berlebih ~ Bila marah, rawit habis ke ahli keluarga ~ Bercakap emosional, merosakkan majlis ~ Melampau dalam bergaduh dan berhujah ~ Jangan bertengkar dan melawak lebih dengan adik beradik ~ Jangan janji dengan adik beradik, tapi tak tunai janji ~ Kecil hati, masak sambal tumis petai, tak mai makan ~ Kadang-kadang kita jaga kawan, lebih dari adik beradik ~ Islam suruh jaga adik beradik dulu ~ Zubair Bin Awwam yang sangat rapat dengan Nabi ~ Jangan bertengkar kalau bab Quran ~ Benda keji dan amalan syaitan yang muktamad ~ Kalau batuk kuat sangat, cuba minum arak secamca ~ Bertengkar dengan orang yang jahil, takkan bawa kebaikan ~ Hadis dalam kitab, tapi palsu ~ Takrij hadis dalam sesuatu kitab ~ Ilmu hadis sangat kompleks ~ Antara kitab-kitab hadis yang kita tak pernah dengar ~ Budak yang akan hembus tempat sujud ~ Jangan sapu debu daripada muka ~ Masalah sapu muka selepas bagi salam ~ Ustaz Shamsuri larang sapu muka ~ Benda kecil, tapi menjadi tanggungjawab untuk semak ~ Jangan jadikan topik besar dalam kuliah di masjid ~ Sesiapa yang menjadikan agama, medan untuk bertengkar ~ Sedih pergaduhan sesama Ustaz ~ Pertengkaran tentang maulid Nabi ~ Bila dibiarkan pertengkaran sampai tak terkawal --- Sokong Projek Zonkuliah Dengan Menyumbang Ke Akaun Berikut : ➢ https://toyyibpay.com/zonkuliah ➢ MAYBANK (Produksi Zonkita) - 557250054584 #Zonkuliah #UstazShamsuri #KuliahAgama

Kumpulan Dakwah Sunnah
Ustadz Abu Yahya Badrusalam, Lc. - Larangan Keras Menerima Perkataan Orang Yang Mengadu-Domba

Kumpulan Dakwah Sunnah

Play Episode Listen Later Jul 30, 2024 57:36


Ustadz Abu Yahya Badrusalam, Lc. - Larangan Keras Menerima Perkataan Orang Yang Mengadu-Domba

The Dissociative Table
Compassion, Curiosity and Collaboration: Kathy Steele's basic treatment approach (Ep. 209)

The Dissociative Table

Play Episode Listen Later Jul 28, 2024 36:02


This episode, "Compassionate Curiosity," features Kathy Steele, famed co-author of The Haunted Self, advocating for a collaborative therapeutic model in resolving severe traumatization, in which attachment phobias are addressed by giving the client (perhaps "gifting" the client) responsibility in treatment. In this manner, the client begins to experience authentic relationship, rather than the therapist supporting maladaptive coping by becoming either a rescuer or an expert (or both). Ms. Steele is a Past President and Fellow of the International Society for the Study of Trauma and Dissociation (ISSTD), and has also served two terms on the Board of the International Society for Traumatic Stress Studies (ISTSS). She has received a number of awards for her work, including the Lifetime Achievement Award from ISSTD, an Emory University Distinguished Alumni Award, and the Cornelia B. Wilbur Award for Outstanding Clinical Contributions from ISSTD. Publications include The Haunted Self (co-author), Treating Trauma-Related Dissociation (first author), and many book chapters. Highlights include: -How therapists facilitate dependency by reinforcing attachment phobias-Polyvagal theory and the collaborative approach -The wisdom of prioritizing relationship over knowing "what's going on"-The importance of the client having a persistent "felt sense" of the therapeutic relationship, rather than having unlimited therapist availability and poor boundaries -Curiosity as a tool for making resistance and relationship rupture the most important moments in treatment-Sharing confusion with the client about treatment direction, to drive collaborative treatment an activate the client's exploration system -Psychoeducational discussion as an underrated toolLook for Ms. Steele's online trainings, resources, books, and consultation groups at kathy-steele.com. In future (as in past) episodes, guests will offer counterpoints to Ms. Steele's views, and this video is a dialogue, rather than an endorsement of her perspective. Likewise, The Dissociative Table does not endorse the views of its hosts, nor do any professional organizations to which the hosts might belong. Join the discussion of this episode and others at the anonymous, free, clinicians-only discussion board by emailing dissociativetable@gmail.com. The music in this episode is taken from the song "Low Winter Sun," from Cracked Machine's album Gates of Keras. The band has generously donated rights to their music in support of increased global access to effective trauma treatment. Stream or download at:https://crackedmachine.bandcamp.com/

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas
280 | François Chollet on Deep Learning and the Meaning of Intelligence

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas

Play Episode Listen Later Jun 24, 2024 101:49


Which is more intelligent, ChatGPT or a 3-year old? Of course this depends on what we mean by "intelligence." A modern LLM is certainly able to answer all sorts of questions that require knowledge far past the capacity of a 3-year old, and even to perform synthetic tasks that seem remarkable to many human grown-ups. But is that really intelligence? François Chollet argues that it is not, and that LLMs are not ever going to be truly "intelligent" in the usual sense -- although other approaches to AI might get there.Support Mindscape on Patreon.Blog post with transcript: https://www.preposterousuniverse.com/podcast/2024/06/24/280-francois-chollet-on-deep-learning-and-the-meaning-of-intelligence/François Chollet received his Diplôme d'Ingénieur from École Nationale Supérieure de Techniques Avancées, Paris. He is currently a Senior Staff Engineer at Google. He has been awarded the Global Swiss AI award for breakthroughs in artificial intelligence. He is the author of Deep Learning with Python, and developer of the Keras software library for neural networks. He is the creator of the ARC (Abstraction and Reasoning Corpus) Challenge.Web siteGithubGoogle Scholar publicationsWikipedia"On the Measure of Intelligence"See Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.

Kencan Dengan Tuhan
Edisi Hari Jumat, 14 Juni 2024 - Bekerja keras sepenuh hati

Kencan Dengan Tuhan

Play Episode Listen Later Jun 13, 2024 5:18


Kencan Dengan Tuhan - Jumat, 14 Juni 2024 Bacaan: "Sebab, juga waktu kami berada di antara kamu, kami memberi peringatan ini kepada kamu: jika seorang tidak mau bekerja, janganlah ia makan." (2 Tesalonika 3:10) Renungan: Tubuh kita terbentuk dari zat-zat yang merupakan karya teragung dari Pencipta kita. Kemampuan kerjanya melebihi segala macam mobil mewah paling otomatis sekalipun. Dengan segala yang kita miliki sebagai manusia, kita bisa menggunakan hidup ini untuk kebaikan atau kejahatan. Kita bisa menggunakannya untuk hidup seperti binatang atau untuk mencapai sukses tertinggi. Tuhan telah membekali kita dengan kekuatan dan kemampuan yang luar biasa. Namun sangat disayangkan bahwa masih banyak manusia yang tak menggunakan berkat Tuhan itu dengan berjuang atau berusaha agar tetap hidup bahkan mencapai hasil yang lebih baik. Bahkan tragisnya, tidak sedikit yang hidupnya tidak mempunyai inisiatif positif. Banyak orang yang hanya menunggu uluran tangan orang lain, padahal ada kemampuan dan kesempatan. Mencari-cari bantuan untuk mempermudah perjalanan hidup dan berfoya-foya tentu tidak layak bagi manusia yang normal karena hidup itu diciptakan Tuhan untuk berusaha. Pada kenyataannya, banyak orang yang kesejahteraannya bisa meningkat karena mereka menggunakan kemampuan dan kesempatan yang Tuhan berikan kepada mereka. Kesejahteraan manusia sesungguhnya dapat diperbaiki jika ia memiliki inisiatif pribadi dan mulai menggunakan tangan serta otak sendiri untuk berusaha. James Garfield seorang mantan Presiden Amerika Serikat dengan bijak mengatakan, "Di dunia ini benda-benda tidak berubah, kecuali kalau ada yang mengubahnya." Perkataan Garfield di atas tentu sangat benar dan itu telah dibuktikannya dalam kehidupannya. Dalam kehidupannya Garfield tak pernah menunggu kesempatan tapi senantiasa berinisiatif menciptakan kesempatan untuk melakukan yang terbaik di dalam hidupnya. Tuhan telah menetapkan manusia untuk mendapatkan rezekinya melalui bekerja dan berpeluh. Hukum itu sampai saat ini masih berlaku karena itu manusia harus berinisiatif untuk bekerja dan berdoa agar ia mendapatkan bagiannya. Zaman semakin susah namun hidup itu harus tetap diperjuangkan karena jika tidak, maka kita akan menderita. Lapangan pekerjaan makin sempit namun bagi orang yang mau berinisiatif, jalan akan dibukakan. Berinisiatiflah selalu di dalam memperjuangkan hidup, pastilah Tuhan akan menolong. Tuhan Yesus memberkati. Doa: Tuhan Yesus, jauhkan dariku kemalasan untuk bekerja dan berusaha serta hanya menggantungkan hidupku pada orang lain. Berikan aku kemauan untuk bekerja dan berkatilah apa saja pekerjaan yang akan aku lakukan. Amin. (Dod).

Radio Rodja 756 AM
Larangan Keras Lewat di Hadapan Orang yang Sedang Shalat

Radio Rodja 756 AM

Play Episode Listen Later Jun 11, 2024


Larangan Keras Lewat di Hadapan Orang yang Sedang Shalat merupakan bagian dari kajian Islam ilmiah Mukhtashar Shahih Muslim yang disampaikan oleh Ustadz Abu Yahya Badrusalam, Lc. Hafidzahullah. Kajian ini disampaikan pada Sabtu, 25 Dzulqa’dah 1445 H / 2 Juni 2024 M. Larangan Keras Lewat di Hadapan Orang yang Sedang Shalat Kita masuk pada bab larangan […] Tulisan Larangan Keras Lewat di Hadapan Orang yang Sedang Shalat ditampilkan di Radio Rodja 756 AM.

כל תכני עושים היסטוריה
מכונות דוברות עברית [עושים תוכנה]

כל תכני עושים היסטוריה

Play Episode Listen Later May 27, 2024 48:50


איך בונים מודלי שפה בעברית? למה זה כל כך מורכב, יקר, ודורש כל כך הרבה מתנדבים? והאם מודלים מקומיים בקוד פתוח יכולים באמת להתחרות במפלצות הבינה המלאכותית של OpenAI וגוגל?בפרק הזה של "עושים תוכנה" מתארח, שוב, ים פלג. לים יש רקורד עשיר מאוד בפיתוח ומחקר סביב AI. הוא מדורג בין המובילים בעולם במדע נתונים תחרותי והחזיק לאורך השנים בכמה שיאי עולם בקטגוריות השונות בתחום. הוא אחד ממפתחיה הראשונים של ספריית הבינה המלאכותית הפופולרית: Keras, ומשמש כמנכ"ל חברת סטארט-אפ שאותה הקים, Deep Trading Ltd, חברת מסחר פיננסי גלובלית, שנשענת על ניסיונו במדעי הנתונים, תכנות ואלגוריתמים. אבל ל-"עושים תוכנה" הוא הגיע הפעם בעיקר כדי לדבר על פרויקט מיוחד, מקומי, שבו הוא מעורב: Ivrit.AI. מיזם ללא מטרות רווח שמטרתו לאפשר תמיכה איכותית בעברית בכלי בינה מלאכותית. אלפי מתנדבים בפרויקט אוספים ומתייגים מאגרי מידע בעברית שלבסוף משוחררים לקהל הרחב ברישיון נוח לשימוש המאפשר לפתח מודלים ושירותי בינה מלאכותית בעברית. המיזם הושק לפני כשנה ואחד היישומים הראשונים שלו הוא מודל תמלול ראשון בעברית, שלפי הדיווחים מתעלה על ביצועי כל המודלים האחרים הקיימים כיום.תפקידו של ים הוא אימון המודל, ויש לו המון דברים מרתקים לספר עליו, על עתיד מודלי השפה, ועל הדרך קדימה עבור הבינה המלאכותית בכלל.האזנה נעימה,בועז לביאלפרק הקודם עם ים:https://www.osimhistoria.com/software/ep123-machine-languageלאתר הפרויקט של Ivrit.AIhttps://www.ivrit.ai/en/ivrit-ai-2/

Change Work Life
A career in data science: what is it, who is it for and how do you get into it? - with Nicholas Cifuentes of WorldQuant University

Change Work Life

Play Episode Listen Later May 7, 2024 48:17 Transcription Available


#179: Nicholas Ciufentes-Goodbody transitioned from a career in teaching to becoming the Chief Data Scientist at WorldQuant University.  He explains what data science involves, what a career in data science looks like, and why it's such a popular field to work in. What you'll learn[1:45] What WorldQuant University is and the types of programs they run. [3:42] How a free university is possible in America.[4:13] The motivation for a hedge fund to run a free university. [4:50] What data science is and how it's used. [6:03] The biggest employers of data scientists. [8:21] What a typical day as a data scientist is like. [10:12] The different specialities within data science. [12:08] What it means to be an AI engineer. [12:40] What qualifications you need to become a data scientist. [14:43] The level of education you need to become a data scientist. [16:27] Character traits that successful data scientists share. [18:10] The amount of nerds working in data science. [20:08] Why musicians become data scientists and doctors. [21:15] How to transition your career to become a data scientist. [27:37] What you need on your resume when applying for a data science role. [29:07] The best ways to learn the skills you need to become a data scientist.  [30:29] How to edit your CV when applying for a data science job. [31:55] How to find a good data science boot camp. [33:51] The base knowledge you need prior to joining a data science boot camp. [36:30] The income potential of a data scientist. [38:17] The career path of a data scientist. [39:28] The impact AI will have on data scientists. [42:12] The ever-changing nature of data science. Resources mentioned in this episodePlease note that some of these are affiliate links and we may get a commission in the event that you make a purchase.  This helps us to cover our expenses and is at no additional cost to you.DataCampKaggleopenAFRICACourseraUdemyAn Introduction to Statistical Learning, Gareth JamesHands-on machine learning with Scikit-Learn, Keras, and TensorFlow, Aurelien GeronFluent Python, Luciano RamalhoMathematics for Human Flourishing, Francis SuFor the show notes for this episode, including a full transcript and links to all the resources mentioned, visit:https://changeworklife.com/using-chatgpt-to-supercharge-your-career/Re-assessing your career?  Know you need a change but don't really know where to start?  Check out these two exercises to start the journey of working out what career is right for you!

People of AI
François Chollet - Creating Keras 3

People of AI

Play Episode Listen Later Apr 18, 2024 65:32


Meet François Chollet, creator of Keras, software engineer, and AI researcher at Google. Join François and hosts Ashley Oldacre and Gus Martins as they discuss how Keras 3 was created, integrating Keras 3 with Gemma and Kaggle, artificial general intelligence (AGI), and much more! Resources: François Chollet research → https://goo.gle/443V3vG Deep Learning With Python, Second Edition → https://goo.gle/3UnpdH1  Intelligence: On Intelligence: How a New Understanding of the Brain Will Lead to the Creation of Truly Intelligent Machines → https://goo.gle/3xDE33s Researcher Pierre-Yves Oudeyer → https://goo.gle/3W8a39V Monty Hall Challenge → https://goo.gle/3VYXAW5  Machine Learning: Keras 3 → https://goo.gle/3JqRgis Gemma on Keras → https://goo.gle/49Q0pfy The ARC challenge on Kaggle → https://goo.gle/3xQsDcr 

Science of Reading Leadership
Superintendent Amy Suman: Leading Literacy Initiatives in a Tribal Community

Science of Reading Leadership

Play Episode Listen Later Mar 14, 2024 14:53


In this episode of SORL, host Dr. Christy Lamb sits down with superintendent Amy Suman to discuss the unique challenges and strategies for literacy achievement in Pueblo of Laguna, New Mexico. Amy shares her dedicated efforts to address equity issues, trauma-informed teaching, and the revitalization of the Keras language. With an emphasis on social determinants and a call to view children as "at promise," Amy's insightful perspective provides valuable lessons in leadership and student support. This episode offers a profound look into the holistic approach to literacy in a close-knit community.Timestamps00:00 Focusing on equity in education for students.04:23 NWEA MAP data shows delayed progress for students.08:52 Reviving tribal language through technology and elders.10:32 Teacher professional development focused on individual coaching.13:17 Importance of spoken language for future literacy.Connect with Amy SumanWebsite: https://www.lagunaed.net/

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

Speaker CFPs and Sponsor Guides are now available for AIE World's Fair — join us on June 25-27 for the biggest AI Engineer conference of 2024!Soumith Chintala needs no introduction in the ML world — his insights are incredibly accessible across Twitter, LinkedIn, podcasts, and conference talks (in this pod we'll assume you'll have caught up on the History of PyTorch pod from last year and cover different topics). He's well known as the creator of PyTorch, but he's more broadly the Engineering Lead on AI Infra, PyTorch, and Generative AI at Meta.Soumith was one of the earliest supporters of Latent Space (and more recently AI News), and we were overjoyed to catch up with him on his latest SF visit for a braindump of the latest AI topics, reactions to some of our past guests, and why Open Source AI is personally so important to him.Life in the GPU-Rich LaneBack in January, Zuck went on Instagram to announce their GPU wealth: by the end of 2024, Meta will have 350k H100s. By adding all their GPU clusters, you'd get to 600k H100-equivalents of compute. At FP16 precision, that's ~1,200,000 PFLOPS. If we used George Hotz's (previous guest!) "Person of Compute" measure, Meta now has 60k humans of compute in their clusters. Occasionally we get glimpses into the GPU-rich life; on a recent ThursdAI chat, swyx prompted PaLM tech lead Yi Tay to write down what he missed most from Google, and he commented that UL2 20B was trained by accidentally leaving the training job running for a month, because hardware failures are so rare in Google.Meta AI's Epic LLM RunBefore Llama broke the internet, Meta released an open source LLM in May 2022, OPT-175B, which was notable for how “open” it was - right down to the logbook! They used only 16 NVIDIA V100 GPUs and Soumith agrees that, with hindsight, it was likely under-trained for its parameter size.In Feb 2023 (pre Latent Space pod), Llama was released, with a 7B version trained on 1T tokens alongside 65B and 33B versions trained on 1.4T tokens. The Llama authors included Guillaume Lample and Timothée Lacroix, who went on to start Mistral.July 2023 was Llama2 time (which we covered!): 3 model sizes, 7B, 13B, and 70B, all trained on 2T tokens. The three models accounted for a grand total of 3,311,616 GPU hours for all pre-training work. CodeLlama followed shortly after, a fine-tune of Llama2 specifically focused on code generation use cases. The family had models in the 7B, 13B, 34B, and 70B size, all trained with 500B extra tokens of code and code-related data, except for 70B which is trained on 1T.All of this on top of other open sourced models like Segment Anything (one of our early hits!), Detectron, Detectron 2, DensePose, and Seamless, and in one year, Meta transformed from a company people made fun of for its “metaverse” investments to one of the key players in the AI landscape and its stock has almost tripled since (about $830B in market value created in the past year).Why Open Source AIThe obvious question is why Meta would spend hundreds of millions on its AI efforts and then release them for free. Zuck has addressed this in public statements:But for Soumith, the motivation is even more personal:“I'm irrationally interested in open source. I think open source has that fundamental way to distribute opportunity in a way that is very powerful. Like, I grew up in India… And knowledge was very centralized, but I saw that evolution of knowledge slowly getting decentralized. And that ended up helping me learn quicker and faster for like zero dollars. And I think that was a strong reason why I ended up where I am. So like that, like the open source side of things, I always push regardless of like what I get paid for, like I think I would do that as a passion project on the side……I think at a fundamental level, the most beneficial value of open source is that you make the distribution to be very wide. It's just available with no friction and people can do transformative things in a way that's very accessible. Maybe it's open source, but it has a commercial license and I'm a student in India. I don't care about the license. I just don't even understand the license. But like the fact that I can use it and do something with it is very transformative to me……Like, okay, I again always go back to like I'm a student in India with no money. What is my accessibility to any of these closed source models? At some scale I have to pay money. That makes it a non-starter and stuff. And there's also the control issue: I strongly believe if you want human aligned AI, you want all humans to give feedback. And you want all humans to have access to that technology in the first place. And I actually have seen, living in New York, whenever I come to Silicon Valley, I see a different cultural bubble.We like the way Soumith put it last year: Closed AI “rate-limits against people's imaginations and needs”!What It Takes For Open Source AI to WinHowever Soumith doesn't think Open Source will simply win by popular demand. There is a tremendous coordination problem with the decentralized nature of the open source AI development right now: nobody is collecting the valuable human feedback in the way that OpenAI or Midjourney are doing.“Open source in general always has a coordination problem. If there's a vertically integrated provider with more resources, they will just be better coordinated than open source. And so now open source has to figure out how to have coordinated benefits. And the reason you want coordinated benefits is because these models are getting better based on human feedback. And if you see with open source models, like if you go to the /r/localllama subreddit, like there's so many variations of models that are being produced from, say, Nous research. I mean, like there's like so many variations built by so many people. And one common theme is they're all using these fine-tuning or human preferences datasets that are very limited and they're not sufficiently diverse. And you look at the other side, say front-ends like Oobabooga or like Hugging Chat or Ollama, they don't really have feedback buttons. All the people using all these front-ends, they probably want to give feedback, but there's no way for them to give feedback… So we're just losing all of this feedback. Maybe open source models are being as used as GPT is at this point in like all kinds of, in a very fragmented way, like in aggregate all the open source models together are probably being used as much as GPT is, maybe close to that. But the amount of feedback that is driving back into the open source ecosystem is like negligible, maybe less than 1% of like the usage. So I think like some, like the blueprint here I think is you'd want someone to create a sinkhole for the feedback… I think if we do that, if that actually happens, I think that probably has a real chance of the open source models having a runaway effect against OpenAI, I think like there's a clear chance we can take at truly winning open source.”If you're working on solving open source coordination, please get in touch!Show Notes* Soumith Chintala Twitter* History of PyTorch episode on Gradient Podcast* The Llama Ecosystem* Apple's MLX* Neural ODEs (Ordinary Differential Equations)* AlphaGo* LMSys arena* Dan Pink's "Drive"* Robotics projects:* Dobb-E* OK Robot* Yann LeCun* Yangqing Jia of Lepton AI* Ed Catmull* George Hotz on Latent Space* Chris Lattner on Latent Space* Guillaume Lample* Yannic Kilcher of OpenAssistant* LMSys* Alex Atallah of OpenRouter* Carlo Sferrazza's 3D tactile research* Alex Wiltschko of Osmo* Tangent by Alex Wiltschko* Lerrel Pinto - RoboticsTimestamps* [00:00:00] Introductions* [00:00:51] Extrinsic vs Intrinsic Success* [00:02:40] Importance of Open Source and Its Impact* [00:03:46] PyTorch vs TinyGrad* [00:08:33] Why PyTorch is the Switzerland of frameworks* [00:10:27] Modular's Mojo + PyTorch?* [00:13:32] PyTorch vs Apple's MLX* [00:16:27] FAIR / PyTorch Alumni* [00:18:50] How can AI inference providers differentiate?* [00:21:41] How to build good benchmarks and learnings from AnyScale's* [00:25:28] Most interesting unexplored ideas* [00:28:18] What people get wrong about synthetic data* [00:35:57] Meta AI's evolution* [00:38:42] How do you allocate 600,000 GPUs?* [00:42:05] Even the GPU Rich are GPU Poor* [00:47:31] Meta's MTIA silicon* [00:50:09] Why we need open source* [00:59:00] Open source's coordination problem for feedback gathering* [01:08:59] Beyond text generation* [01:15:37] Osmo and the Future of Smell Recognition TechnologyTranscriptAlessio [00:00:00]: Hey everyone, welcome to the Latent Space podcast. This is Alessio, partner and CTO in residence at Decibel Partners, and I'm joined by my co-host Swyx, founder of Smol AI.Swyx [00:00:15]: Hey, and today we have in the studio Soumith Chintala, welcome.Soumith [00:00:17]: Thanks for having me.Swyx [00:00:18]: On one of your rare visits from New York where you live. You got your start in computer vision at NYU with Yann LeCun. That was a very fortuitous start. I was actually listening to your interview on the Gradient podcast. So if people want to know more about the history of Soumith, history of PyTorch, they can go to that podcast. We won't spend that much time there, but I just was marveling at your luck, or I don't know if it's your luck or your drive to find AI early and then find the right quality mentor because I guess Yan really sort of introduced you to that world.Soumith [00:00:51]: Yeah, I think you're talking about extrinsic success, right? A lot of people just have drive to do things that they think is fun, and a lot of those things might or might not be extrinsically perceived as good and successful. I think I just happened to like something that is now one of the coolest things in the world or whatever. But if I happen, the first thing I tried to become was a 3D VFX artist, and I was really interested in doing that, but I turned out to be very bad at it. So I ended up not doing that further. But even if I was good at that, whatever, and I ended up going down that path, I probably would have been equally happy. It's just like maybe like the perception of, oh, is this person successful or not might be different. I think like after a baseline, like your happiness is probably more correlated with your intrinsic stuff.Swyx [00:01:44]: Yes. I think Dan Pink has this book on drive that I often refer to about the power of intrinsic motivation versus extrinsic and how long extrinsic lasts. It's not very long at all. But anyway, now you are an investor in Runway, so in a way you're working on VFX. Yes.Soumith [00:02:01]: I mean, in a very convoluted way.Swyx [00:02:03]: It reminds me of Ed Catmull. I don't know if you guys know, but he actually tried to become an animator in his early years and failed or didn't get accepted by Disney and then went and created Pixar and then got bought by Disney and created Toy Story. So you joined Facebook in 2014 and eventually became a creator and maintainer of PyTorch. And there's this long story there you can refer to on the gradient. I think maybe people don't know that you also involved in more sort of hardware and cluster decision affair. And we can dive into more details there because we're all about hardware this month. Yeah. And then finally, I don't know what else, like what else should people know about you on a personal side or professional side?Soumith [00:02:40]: I think open source is definitely a big passion of mine and probably forms a little bit of my identity at this point. I'm irrationally interested in open source. I think open source has that fundamental way to distribute opportunity in a way that is very powerful. Like, I grew up in India. I didn't have internet for a while. In college, actually, I didn't have internet except for GPRS or whatever. And knowledge was very centralized, but I saw that evolution of knowledge slowly getting decentralized. And that ended up helping me learn quicker and faster for zero dollars. And I think that was a strong reason why I ended up where I am. So the open source side of things, I always push regardless of what I get paid for, like I think I would do that as a passion project on the side.Swyx [00:03:35]: Yeah, that's wonderful. Well, we'll talk about the challenges as well that open source has, open models versus closed models. Maybe you want to touch a little bit on PyTorch before we move on to the sort of Meta AI in general.PyTorch vs Tinygrad tradeoffsAlessio [00:03:46]: Yeah, we kind of touched on PyTorch in a lot of episodes. So we had George Hotz from TinyGrad. He called PyTorch a CISC and TinyGrad a RISC. I would love to get your thoughts on PyTorch design direction as far as, I know you talk a lot about kind of having a happy path to start with and then making complexity hidden away but then available to the end user. One of the things that George mentioned is I think you have like 250 primitive operators in PyTorch, I think TinyGrad is four. So how do you think about some of the learnings that maybe he's going to run into that you already had in the past seven, eight years almost of running PyTorch?Soumith [00:04:24]: Yeah, I think there's different models here, but I think it's two different models that people generally start with. Either they go like, I have a grand vision and I'm going to build a giant system that achieves this grand vision and maybe one is super feature complete or whatever. Or other people say they will get incrementally ambitious, right? And they say, oh, we'll start with something simple and then we'll slowly layer out complexity in a way that optimally applies Huffman coding or whatever. Like where the density of users are and what they're using, I would want to keep it in the easy, happy path and where the more niche advanced use cases, I'll still want people to try them, but they need to take additional frictional steps. George, I think just like we started with PyTorch, George started with the incrementally ambitious thing. I remember TinyGrad used to be, like we would be limited to a thousand lines of code and I think now it's at 5,000. So I think there is no real magic to which why PyTorch has the kind of complexity. I think it's probably partly necessitated and partly because we built with the technology available under us at that time, PyTorch is like 190,000 lines of code or something at this point. I think if you had to rewrite it, we would probably think about ways to rewrite it in a vastly simplified way for sure. But a lot of that complexity comes from the fact that in a very simple, explainable way, you have memory hierarchies. You have CPU has three levels of caches and then you have DRAM and SSD and then you have network. Similarly, GPU has several levels of memory and then you have different levels of network hierarchies, NVLink plus InfiniBand or Rocky or something like that, right? And the way the flops are available on your hardware, they are available in a certain way and your computation is in a certain way and you have to retrofit your computation onto both the memory hierarchy and like the flops available. When you're doing this, it is actually a fairly hard mathematical problem to do this setup, like you find the optimal thing. And finding the optimal thing is, what is optimal depends on the input variables themselves. So like, okay, what is the shape of your input tensors and what is the operation you're trying to do and various things like that. Finding that optimal configuration and writing it down in code is not the same for every input configuration you have. Like for example, just as the shape of the tensors change, let's say you have three input tensors into a Sparstar product or something like that. The shape of each of these input tensors will vastly change how you do this optimally placing this operation onto the hardware in a way that will get you maximal throughput. So a lot of our complexity comes from writing out hundreds of configurations for each single PyTorch operator and templatizing these things and symbolically generating the final CUDA code or CPU code. There's no way to avoid it because mathematically we haven't found symbolic ways to do this that also keep compile time near zero. You can write a very simple framework, but then you also should be willing to eat the long compile time. So if searching for that optimal performance at runtime, but that's the trade off. There's no, like, I don't think unless we have great breakthroughs George's vision is achievable, he should be thinking about a narrower problem such as I'm only going to make this for work for self-driving car connets or I'm only going to make this work for LLM transformers of the llama style. Like if you start narrowing the problem down, you can make a vastly simpler framework. But if you don't, if you need the generality to power all of the AI research that is happening and keep zero compile time and in all these other factors, I think it's not easy to avoid the complexity.Pytorch vs MojoAlessio [00:08:33]: That's interesting. And we kind of touched on this with Chris Lattner when he was on the podcast. If you think about frameworks, they have the model target. They have the hardware target. They have different things to think about. He mentioned when he was at Google, TensorFlow trying to be optimized to make TPUs go brr, you know, and go as fast. I think George is trying to make especially AMD stack be better than ROCm. How come PyTorch has been such as Switzerland versus just making Meta hardware go brr?Soumith [00:09:00]: First, Meta is not in the business of selling hardware. Meta is not in the business of cloud compute. The way Meta thinks about funding PyTorch is we're funding it because it's net good for Meta to fund PyTorch because PyTorch has become a standard and a big open source project. And generally it gives us a timeline edge. It gives us leverage and all that within our own work. So why is PyTorch more of a Switzerland rather than being opinionated? I think the way we think about it is not in terms of Switzerland or not. We actually the way we articulate it to all hardware vendors and software vendors and all who come to us being we want to build a backend in core for PyTorch and ship it by default is we just only look at our user side of things. Like if users are using a particular piece of hardware, then we want to support it. We very much don't want to king make the hardware side of things. So as the MacBooks have GPUs and as that stuff started getting increasingly interesting, we pushed Apple to push some engineers and work on the NPS support and we spend significant time from Meta funded engineers on that as well because a lot of people are using the Apple GPUs and there's demand. So we kind of mostly look at it from the demand side. We never look at it from like oh which hardware should we start taking opinions on.Swyx [00:10:27]: Is there a future in which, because Mojo or Modular Mojo is kind of a superset of Python, is there a future in which PyTorch might use Mojo features optionally?Soumith [00:10:36]: I think it depends on how well integrated it is into the Python ecosystem. So if Mojo is like a pip install and it's readily available and users feel like they can use Mojo so smoothly within their workflows in a way that just is low friction, we would definitely look into that. Like in the same way PyTorch now depends on Triton, OpenAI Triton, and we never had a conversation that was like huh, that's like a dependency. Should we just build a Triton of our own or should we use Triton? It almost doesn't, like those conversations don't really come up for us. The conversations are more well does Triton have 10,000 dependencies and is it hard to install? We almost don't look at these things from a strategic leverage point of view. We look at these things from a user experience point of view, like is it easy to install? Is it smoothly integrated and does it give enough benefits for us to start depending on it? If so, yeah, we should consider it. That's how we think about it.Swyx [00:11:37]: You're inclusive by default as long as it meets the minimum bar of, yeah, but like maybe I phrased it wrongly. Maybe it's more like what problems would you look to solve that you have right now?Soumith [00:11:48]: I think it depends on what problems Mojo will be useful at.Swyx [00:11:52]: Mainly a performance pitch, some amount of cross compiling pitch.Soumith [00:11:56]: Yeah, I think the performance pitch for Mojo was like, we're going to be performant even if you have a lot of custom stuff, you're going to write arbitrary custom things and we will be performant. And that value proposition is not clear to us from the PyTorch side to consider it for PyTorch. So PyTorch, it's actually not 250 operators, it's like a thousand operators. PyTorch exposes about a thousand operators and people kind of write their ideas in the thousand operators of PyTorch. Mojo is like, well, maybe it's okay to completely sidestep those thousand operators of PyTorch and just write it in a more natural form. Just write raw Python, write for loops or whatever, right? So from the consideration of how do we intersect PyTorch with Mojo, I can see one use case where you have custom stuff for some parts of your program, but mostly it's PyTorch. And so we can probably figure out how to make it easier for say Torch.compile to smoothly also consume Mojo subgraphs and like, you know, the interoperability being actually usable, that I think is valuable. But Mojo as a fundamental front end would be replacing PyTorch, not augmenting PyTorch. So in that sense, I don't see a synergy in more deeply integrating Mojo.Pytorch vs MLXSwyx [00:13:21]: So call out to Mojo whenever they have written something in Mojo and there's some performance related thing going on. And then since you mentioned Apple, what should people think of PyTorch versus MLX?Soumith [00:13:32]: I mean, MLX is early and I know the folks well, Ani used to work at FAIR and I used to chat with him all the time. He used to be based out of New York as well. The way I think about MLX is that MLX is specialized for Apple right now. It has a happy path because it's defined its product in a narrow way. At some point MLX either says we will only be supporting Apple and we will just focus on enabling, you know, there's a framework if you use your MacBook, but once you like go server side or whatever, that's not my problem and I don't care. For MLS, it enters like the server side set of things as well. Like one of these two things will happen, right? If the first thing will happen, like MLX's overall addressable market will be small, but it probably do well within that addressable market. If it enters the second phase, they're going to run into all the same complexities that we have to deal with. They will not have any magic wand and they will have more complex work to do. They probably wouldn't be able to move as fast.Swyx [00:14:44]: Like having to deal with distributed compute?Soumith [00:14:48]: Distributed, NVIDIA and AMD GPUs, like just like having a generalization of the concept of a backend, how they treat compilation with plus overheads. Right now they're deeply assumed like the whole NPS graph thing. So they need to think about all these additional things if they end up expanding onto the server side and they'll probably build something like PyTorch as well, right? Like eventually that's where it will land. And I think there they will kind of fail on the lack of differentiation. Like it wouldn't be obvious to people why they would want to use it.Swyx [00:15:24]: I mean, there are some cloud companies offering M1 and M2 chips on servers. I feel like it might be interesting for Apple to pursue that market, but it's not their core strength.Soumith [00:15:33]: Yeah. If Apple can figure out their interconnect story, maybe, like then it can become a thing.Swyx [00:15:40]: Honestly, that's more interesting than the cars. Yes.Soumith [00:15:43]: I think the moat that NVIDIA has right now, I feel is that they have the interconnect that no one else has, like AMD GPUs are pretty good. I'm sure there's various silicon that is not bad at all, but the interconnect, like NVLink is uniquely awesome. I'm sure the other hardware providers are working on it, but-Swyx [00:16:04]: I feel like when you say it's uniquely awesome, you have some appreciation of it that the rest of us don't. I mean, the rest of us just like, you know, we hear marketing lines, but what do you mean when you say NVIDIA is very good at networking? Obviously they made the acquisition maybe like 15 years ago.Soumith [00:16:15]: Just the bandwidth it offers and the latency it offers. I mean, TPUs also have a good interconnect, but you can't buy them. So you have to go to Google to use it.PyTorch MafiaAlessio [00:16:27]: Who are some of the other FAIR PyTorch alumni that are building cool companies? I know you have Fireworks AI, Lightning AI, Lepton, and Yangqing, you knew since college when he was building Coffee?Soumith [00:16:40]: Yeah, so Yangqing and I used to be framework rivals, PyTorch, I mean, we were all a very small close-knit community back then. Caffe, Torch, Theano, Chainer, Keras, various frameworks. I mean, it used to be more like 20 frameworks. I can't remember all the names. CCV by Liu Liu, who is also based out of SF. And I would actually like, you know, one of the ways it was interesting is you went into the framework guts and saw if someone wrote their own convolution kernel or they were just copying someone else's. There were four or five convolution kernels that were unique and interesting. There was one from this guy out of Russia, I forgot the name, but I remembered who was awesome enough to have written their own kernel. And at some point there, I built out these benchmarks called ConNet benchmarks. They're just benchmarking all the convolution kernels that are available at that time. It hilariously became big enough that at that time AI was getting important, but not important enough that industrial strength players came in to do these kinds of benchmarking and standardization. Like we have MLPerf today. So a lot of the startups were using ConNet benchmarks in their pitch decks as like, oh, you know, on ConNet benchmarks, this is how we fare, so you should fund us. I remember Nirvana actually was at the top of the pack because Scott Gray wrote amazingly fast convolution kernels at that time. Very interesting, but separate times. But to answer your question, Alessio, I think mainly Lepton, Fireworks are the two most obvious ones, but I'm sure the fingerprints are a lot wider. They're just people who worked within the PyTorch Cafe2 cohort of things and now end up at various other places.Swyx [00:18:50]: I think as a, both as an investor and a people looking to build on top of their services, it's a uncomfortable slash like, I don't know what I don't know pitch. Because I've met Yang Tsing and I've met Lin Chao. Yeah, I've met these folks and they're like, you know, we are deep in the PyTorch ecosystem and we serve billions of inferences a day or whatever at Facebook and now we can do it for you. And I'm like, okay, that's great. Like, what should I be wary of or cautious of when these things happen? Because I'm like, obviously this experience is extremely powerful and valuable. I just don't know what I don't know. Like, what should people know about like these sort of new inference as a service companies?Soumith [00:19:32]: I think at that point you would be investing in them for their expertise of one kind. So if they've been at a large company, but they've been doing amazing work, you would be thinking about it as what these people bring to the table is that they're really good at like GPU programming or understanding the complexity of serving models once it hits a certain scale. You know, various expertise like from the infra and AI and GPUs point of view. What you would obviously want to figure out is whether their understanding of the external markets is clear, whether they know and understand how to think about running a business, understanding how to be disciplined about making money or, you know, various things like that.Swyx [00:20:23]: Maybe I'll put it like, actually I will de-emphasize the investing bit and just more as a potential customer. Oh, okay. Like, it's more okay, you know, you have PyTorch gods, of course. Like, what else should I know?Soumith [00:20:37]: I mean, I would not care about who's building something. If I'm trying to be a customer, I would care about whether...Swyx [00:20:44]: Benchmarks.Soumith [00:20:44]: Yeah, I use it and it's usability and reliability and speed, right?Swyx [00:20:51]: Quality as well.Soumith [00:20:51]: Yeah, if someone from some random unknown place came to me and say, user stuff is great. Like, and I have the bandwidth, I probably will give it a shot. And if it turns out to be great, like I'll just use it.Benchmark dramaSwyx [00:21:07]: Okay, great. And then maybe one more thing about benchmarks, since we already brought it up and you brought up Confident Benchmarks. There was some recent drama around AnyScale. AnyScale released their own benchmarks and obviously they look great on their own benchmarks, but maybe didn't give the other... I feel there are two lines of criticism. One, which is they didn't test some apples for apples on the kind of endpoints that the other providers, that they are competitors with, on their benchmarks and that is due diligence baseline. And then the second would be more just optimizing for the right thing. You had some commentary on it. I'll just kind of let you riff.Soumith [00:21:41]: Yeah, I mean, in summary, basically my criticism of that was AnyScale built these benchmarks for end users to just understand what they should pick, right? And that's a very good thing to do. I think what they didn't do a good job of is give that end user a full understanding of what they should pick. Like they just gave them a very narrow slice of understanding. I think they just gave them latency numbers and that's not sufficient, right? You need to understand your total cost of ownership at some reasonable scale. Not oh, one API call is one cent, but a thousand API calls are 10 cents. Like people can misprice to cheat on those benchmarks. So you want to understand, okay, like how much is it going to cost me if I actually subscribe to you and do like a million API calls a month or something? And then you want to understand the latency and reliability, not just from one call you made, but an aggregate of calls you've made over several various times of the day and times of the week. And the nature of the workloads, is it just some generic single paragraph that you're sending that is cashable? Or is it like testing of real world workload? I think that kind of rigor, like in presenting that benchmark wasn't there. It was a much more narrow sliver of what should have been a good benchmark. That was my main criticism. And I'm pretty sure if before they released it, they showed it to their other stakeholders who would be caring about this benchmark because they are present in it, they would have easily just pointed out these gaps. And I think they didn't do that and they just released it. So I think those were the two main criticisms. I think they were fair and Robert took it well.Swyx [00:23:40]: And he took it very well. And we'll have him on at some point and we'll discuss it. But I think it's important for, I think the market being maturing enough that people start caring and competing on these kinds of things means that we need to establish what best practice is because otherwise everyone's going to play dirty.Soumith [00:23:55]: Yeah, absolutely. My view of the LLM inference market in general is that it's the laundromat model. Like the margins are going to drive down towards the bare minimum. It's going to be all kinds of arbitrage between how much you can get the hardware for and then how much you sell the API and how much latency your customers are willing to let go. You need to figure out how to squeeze your margins. Like what is your unique thing here? Like I think Together and Fireworks and all these people are trying to build some faster CUDA kernels and faster, you know, hardware kernels in general. But those modes only last for a month or two. These ideas quickly propagate.Swyx [00:24:38]: Even if they're not published?Soumith [00:24:39]: Even if they're not published, the idea space is small. So even if they're not published, the discovery rate is going to be pretty high. It's not like we're talking about a combinatorial thing that is really large. You're talking about Llama style LLM models. And we're going to beat those to death on a few different hardware SKUs, right? Like it's not even we have a huge diversity of hardware you're going to aim to run it on. Now when you have such a narrow problem and you have a lot of people working on it, the rate at which these ideas are going to get figured out is going to be pretty rapid.Swyx [00:25:15]: Is it a standard bag of tricks? Like the standard one that I know of is, you know, fusing operators and-Soumith [00:25:22]: Yeah, it's the standard bag of tricks on figuring out how to improve your memory bandwidth and all that, yeah.Alessio [00:25:28]: Any ideas instead of things that are not being beaten to death that people should be paying more attention to?Novel PyTorch ApplicationsSwyx [00:25:34]: One thing I was like, you know, you have a thousand operators, right? Like what's the most interesting usage of PyTorch that you're seeing maybe outside of this little bubble?Soumith [00:25:41]: So PyTorch, it's very interesting and scary at the same time, but basically it's used in a lot of exotic ways, like from the ML angle, what kind of models are being built? And you get all the way from state-based models and all of these things to stuff nth order differentiable models, like neural ODEs and stuff like that. I think there's one set of interestingness factor from the ML side of things. And then there's the other set of interesting factor from the applications point of view. It's used in Mars Rover simulations, to drug discovery, to Tesla cars. And there's a huge diversity of applications in which it is used. So in terms of the most interesting application side of things, I think I'm scared at how many interesting things that are also very critical and really important it is used in. I think the scariest was when I went to visit CERN at some point and they said they were using PyTorch and they were using GANs at the same time for particle physics research. And I was scared more about the fact that they were using GANs than they were using PyTorch, because at that time I was a researcher focusing on GANs. But the diversity is probably the most interesting. How many different things it is being used in. I think that's the most interesting to me from the applications perspective. From the models perspective, I think I've seen a lot of them. Like the really interesting ones to me are where we're starting to combine search and symbolic stuff with differentiable models, like the whole AlphaGo style models is one example. And then I think we're attempting to do it for LLMs as well, with various reward models and search. I mean, I don't think PyTorch is being used in this, but the whole alpha geometry thing was interesting because again, it's an example of combining the symbolic models with the gradient based ones. But there are stuff like alpha geometry that PyTorch is used at, especially when you intersect biology and chemistry with ML. In those areas, you want stronger guarantees on the output. So yeah, maybe from the ML side, those things to me are very interesting right now.Swyx [00:28:03]: Yeah. People are very excited about the alpha geometry thing. And it's kind of like, for me, it's theoretical. It's great. You can solve some Olympia questions. I'm not sure how to make that bridge over into the real world applications, but I'm sure people smarter than me will figure it out.Synthetic Data vs Symbolic ModelsSoumith [00:28:18]: Let me give you an example of it. You know how the whole thing about synthetic data will be the next rage in LLMs is a thing?Swyx [00:28:27]: Already is a rage.Soumith [00:28:28]: Which I think is fairly misplaced in how people perceive it. People think synthetic data is some kind of magic wand that you wave and it's going to be amazing. Synthetic data is useful in neural networks right now because we as humans have figured out a bunch of symbolic models of the world or made up certain symbolic models because of human innate biases. So we've figured out how to ground particle physics in a 30 parameter model. And it's just very hard to compute as in it takes a lot of flops to compute, but it only has 30 parameters or so. I mean, I'm not a physics expert, but it's a very low rank model. We built mathematics as a field that basically is very low rank. Language, a deep understanding of language, like the whole syntactic parse trees and just understanding how language can be broken down and into a formal symbolism is something that we figured out. So we basically as humans have accumulated all this knowledge on these subjects, either synthetic, we created those subjects in our heads, or we grounded some real world phenomenon into a set of symbols. But we haven't figured out how to teach neural networks symbolic world models directly. The only way we have to teach them is generating a bunch of inputs and outputs and gradient dissenting over them. So in areas where we have the symbolic models and we need to teach all the knowledge we have that is better encoded in the symbolic models, what we're doing is we're generating a bunch of synthetic data, a bunch of input output pairs, and then giving that to the neural network and asking it to learn the same thing that we already have a better low rank model of in gradient descent in a much more over-parameterized way. Outside of this, like where we don't have good symbolic models, like synthetic data obviously doesn't make any sense. So synthetic data is not a magic wand where it'll work in all cases in every case or whatever. It's just where we as humans already have good symbolic models off. We need to impart that knowledge to neural networks and we figured out the synthetic data is a vehicle to impart this knowledge to. So, but people, because maybe they don't know enough about synthetic data as a notion, but they hear, you know, the next wave of data revolution is synthetic data. They think it's some kind of magic where we just create a bunch of random data somehow. They don't think about how, and then they think that's just a revolution. And I think that's maybe a gap in understanding most people have in this hype cycle.Swyx [00:31:23]: Yeah, well, it's a relatively new concept, so. Oh, there's two more that I'll put in front of you and then you can see what you respond. One is, you know, I have this joke that it's, you know, it's only synthetic data if it's from the Mistral region of France, otherwise it's just a sparkling distillation, which is what news research is doing. Like they're distilling GPT-4 by creating synthetic data from GPT-4, creating mock textbooks inspired by Phi 2 and then fine tuning open source models like Llama. And so I don't know, I mean, I think that's, should we call that synthetic data? Should we call it something else? I don't know.Soumith [00:31:57]: Yeah, I mean, the outputs of LLMs, are they synthetic data? They probably are, but I think it depends on the goal you have. If your goal is you're creating synthetic data with the goal of trying to distill GPT-4's superiority into another model, I guess you can call it synthetic data, but it also feels like disingenuous because your goal is I need to copy the behavior of GPT-4 and-Swyx [00:32:25]: It's also not just behavior, but data set. So I've often thought of this as data set washing. Like you need one model at the top of the chain, you know, unnamed French company that has that, you know, makes a model that has all the data in it that we don't know where it's from, but it's open source, hey, and then we distill from that and it's great. To be fair, they also use larger models as judges for preference ranking, right? So that is, I think, a very, very accepted use of synthetic.Soumith [00:32:53]: Correct. I think it's a very interesting time where we don't really have good social models of what is acceptable depending on how many bits of information you use from someone else, right? It's like, okay, you use one bit. Is that okay? Yeah, let's accept it to be okay. Okay, what about if you use 20 bits? Is that okay? I don't know. What if you use 200 bits? I don't think we as society have ever been in this conundrum where we have to be like, where is the boundary of copyright or where is the boundary of socially accepted understanding of copying someone else? We haven't been tested this mathematically before,Swyx [00:33:38]: in my opinion. Whether it's transformative use. Yes. So yeah, I think this New York Times opening eye case is gonna go to the Supreme Court and we'll have to decide it because I think we never had to deal with it before. And then finally, for synthetic data, the thing that I'm personally exploring is solving this great stark paradigm difference between rag and fine tuning, where you can kind of create synthetic data off of your retrieved documents and then fine tune on that. That's kind of synthetic. All you need is variation or diversity of samples for you to fine tune on. And then you can fine tune new knowledge into your model. I don't know if you've seen that as a direction for synthetic data.Soumith [00:34:13]: I think you're basically trying to, what you're doing is you're saying, well, language, I know how to parametrize language to an extent. And I need to teach my model variations of this input data so that it's resilient or invariant to language uses of that data.Swyx [00:34:32]: Yeah, it doesn't overfit on the wrong source documents.Soumith [00:34:33]: So I think that's 100% synthetic. You understand, the key is you create variations of your documents and you know how to do that because you have a symbolic model or like some implicit symbolic model of language.Swyx [00:34:48]: Okay.Alessio [00:34:49]: Do you think the issue with symbolic models is just the architecture of the language models that we're building? I think maybe the thing that people grasp is the inability of transformers to deal with numbers because of the tokenizer. Is it a fundamental issue there too? And do you see alternative architectures that will be better with symbolic understanding?Soumith [00:35:09]: I am not sure if it's a fundamental issue or not. I think we just don't understand transformers enough. I don't even mean transformers as an architecture. I mean the use of transformers today, like combining the tokenizer and transformers and the dynamics of training, when you show math heavy questions versus not. I don't have a good calibration of whether I know the answer or not. I, you know, there's common criticisms that are, you know, transformers will just fail at X. But then when you scale them up to sufficient scale, they actually don't fail at that X. I think there's this entire subfield where they're trying to figure out these answers called like the science of deep learning or something. So we'll get to know more. I don't know the answer.Meta AI and Llama 2/3Swyx [00:35:57]: Got it. Let's touch a little bit on just Meta AI and you know, stuff that's going on there. Maybe, I don't know how deeply you're personally involved in it, but you're our first guest with Meta AI, which is really fantastic. And Llama 1 was, you know, you are such a believer in open source. Llama 1 was more or less the real breakthrough in open source AI. The most interesting thing for us covering on this, in this podcast was the death of Chinchilla, as people say. Any interesting insights there around the scaling models for open source models or smaller models or whatever that design decision was when you guys were doing it?Soumith [00:36:31]: So Llama 1 was Guillaume Lample and team. There was OPT before, which I think I'm also very proud of because we bridged the gap in understanding of how complex it is to train these models to the world. Like until then, no one really in gory detail published.Swyx [00:36:50]: The logs.Soumith [00:36:51]: Yeah. Like, why is it complex? And everyone says, oh, it's complex. But no one really talked about why it's complex. I think OPT was cool.Swyx [00:37:02]: I met Susan and she's very, very outspoken. Yeah.Soumith [00:37:05]: We probably, I think, didn't train it for long enough, right? That's kind of obvious in retrospect.Swyx [00:37:12]: For a 175B. Yeah. You trained it according to Chinchilla at the time or?Soumith [00:37:17]: I can't remember the details, but I think it's a commonly held belief at this point that if we trained OPT longer, it would actually end up being better. Llama 1, I think, was Guillaume Lample and team Guillaume is fantastic and went on to build Mistral. I wasn't too involved in that side of things. So I don't know what you're asking me, which is how did they think about scaling loss and all of that? Llama 2, I was more closely involved in. I helped them a reasonable amount with their infrastructure needs and stuff. And Llama 2, I think, was more like, let's get to the evolution. At that point, we kind of understood what we were missing from the industry's understanding of LLMs. And we needed more data and we needed more to train the models for longer. And we made, I think, a few tweaks to the architecture and we scaled up more. And that was Llama 2. I think Llama 2, you can think of it as after Guillaume left, the team kind of rebuilt their muscle around Llama 2. And Hugo, I think, who's the first author is fantastic. And I think he did play a reasonable big role in Llama 1 as well.Soumith [00:38:35]: And he overlaps between Llama 1 and 2. So in Llama 3, obviously, hopefully, it'll be awesome.Alessio [00:38:42]: Just one question on Llama 2, and then we'll try and fish Llama 3 spoilers out of you. In the Llama 2 paper, the loss curves of the 34 and 70B parameter, they still seem kind of steep. Like they could go lower. How, from an infrastructure level, how do you allocate resources? Could they have just gone longer or were you just, hey, this is all the GPUs that we can burn and let's just move on to Llama 3 and then make that one better?Soumith [00:39:07]: Instead of answering specifically about that Llama 2 situation or whatever, I'll tell you how we think about things. Generally, we're, I mean, Mark really is some numbers, right?Swyx [00:39:20]: So let's cite those things again. All I remember is like 600K GPUs.Soumith [00:39:24]: That is by the end of this year and 600K H100 equivalents. With 250K H100s, including all of our other GPU or accelerator stuff, it would be 600-and-something-K aggregate capacity.Swyx [00:39:38]: That's a lot of GPUs.Soumith [00:39:39]: We'll talk about that separately. But the way we think about it is we have a train of models, right? Llama 1, 2, 3, 4. And we have a bunch of GPUs. I don't think we're short of GPUs. Like-Swyx [00:39:54]: Yeah, no, I wouldn't say so. Yeah, so it's all a matter of time.Soumith [00:39:56]: I think time is the biggest bottleneck. It's like, when do you stop training the previous one and when do you start training the next one? And how do you make those decisions? The data, do you have net new data, better clean data for the next one in a way that it's not worth really focusing on the previous one? It's just a standard iterative product. You're like, when is the iPhone 1? When do you start working on iPhone 2? Where is the iPhone? And so on, right? So mostly the considerations are time and generation, rather than GPUs, in my opinion.Alessio [00:40:31]: So one of the things with the scaling loss, like Chinchilla is optimal to balance training and inference costs. I think at Meta's scale, you would rather pay a lot more maybe at training and then save on inference. How do you think about that from infrastructure perspective? I think in your tweet, you say you can try and guess on like how we're using these GPUs. Can you just give people a bit of understanding? It's like, because I've already seen a lot of VCs say, Llama 3 has been trained on 600,000 GPUs and that's obviously not true, I'm sure. How do you allocate between the research, FAIR and the Llama training, the inference on Instagram suggestions that get me to scroll, like AI-generated stickers on WhatsApp and all of that?Soumith [00:41:11]: Yeah, we haven't talked about any of this publicly, but as a broad stroke, it's like how we would allocate resources of any other kinds at any company. You run a VC portfolio, how do you allocate your investments between different companies or whatever? You kind of make various trade-offs and you kind of decide, should I invest in this project or this other project, or how much should I invest in this project? It's very much a zero sum of trade-offs. And it also comes into play, how are your clusters configured, like overall, what you can fit of what size and what cluster and so on. So broadly, there's no magic sauce here. I mean, I think the details would add more spice, but also wouldn't add more understanding. It's just gonna be like, oh, okay, I mean, this looks like they just think about this as I would normally do.Alessio [00:42:05]: So even the GPU rich run through the same struggles of having to decide where to allocate things.Soumith [00:42:11]: Yeah, I mean, at some point I forgot who said it, but you kind of fit your models to the amount of compute you have. If you don't have enough compute, you figure out how to make do with smaller models. But no one as of today, I think would feel like they have enough compute. I don't think I've heard any company within the AI space be like, oh yeah, like we feel like we have sufficient compute and we couldn't have done better. So that conversation, I don't think I've heard from any of my friends at other companies.EleutherSwyx [00:42:47]: Stella from Eleuther sometimes says that because she has a lot of donated compute. She's trying to put it to interesting uses, but for some reason she's decided to stop making large models.Soumith [00:42:57]: I mean, that's a cool, high conviction opinion that might pay out.Swyx [00:43:01]: Why?Soumith [00:43:02]: I mean, she's taking a path that most people don't care to take about in this climate and she probably will have very differentiated ideas. I mean, think about the correlation of ideas in AI right now. It's so bad, right? So everyone's fighting for the same pie. In some weird sense, that's partly why I don't really directly work on LLMs. I used to do image models and stuff and I actually stopped doing GANs because GANs were getting so hot that I didn't have any calibration of whether my work would be useful or not because, oh yeah, someone else did the same thing you did. It's like, there's so much to do, I don't understand why I need to fight for the same pie. So I think Stella's decision is very smart.Making BetsAlessio [00:43:53]: And how do you reconcile that with how we started the discussion about intrinsic versus extrinsic kind of like accomplishment or success? How should people think about that especially when they're doing a PhD or early in their career? I think in Europe, I walked through a lot of the posters and whatnot, there seems to be mode collapse in a way in the research, a lot of people working on the same things. Is it worth for a PhD to not take a bet on something that is maybe not as interesting just because of funding and visibility and whatnot? Or yeah, what suggestions would you give?Soumith [00:44:28]: I think there's a baseline level of compatibility you need to have with the field. Basically, you need to figure out if you will get paid enough to eat, right? Like whatever reasonable normal lifestyle you want to have as a baseline. So you at least have to pick a problem within the neighborhood of fundable. Like you wouldn't wanna be doing something so obscure that people are like, I don't know, like you can work on it.Swyx [00:44:59]: Would a limit on fundability, I'm just observing something like three months of compute, right? That's the top line, that's the like max that you can spend on any one project.Soumith [00:45:09]: But like, I think that's very ill specified, like how much compute, right? I think that the notion of fundability is broader. It's more like, hey, are these family of models within the acceptable set of, you're not crazy or something, right? Even something like neural or DS, which is a very boundary pushing thing or states-based models or whatever. Like all of these things I think are still in fundable territory. When you're talking about, I'm gonna do one of the neuromorphic models and then apply image classification to them or something, then it becomes a bit questionable. Again, it depends on your motivation. Maybe if you're a neuroscientist, it actually is feasible. But if you're an AI engineer, like the audience of these podcasts, then it's more questionable. The way I think about it is, you need to figure out how you can be in the baseline level of fundability just so that you can just live. And then after that, really focus on intrinsic motivation and depends on your strengths, like how you can play to your strengths and your interests at the same time. Like I try to look at a bunch of ideas that are interesting to me, but also try to play to my strengths. I'm not gonna go work on theoretical ML. I'm interested in it, but when I want to work on something like that, I try to partner with someone who is actually a good theoretical ML person and see if I actually have any value to provide. And if they think I do, then I come in. So I think you'd want to find that intersection of ideas you like, and that also play to your strengths. And I'd go from there. Everything else, like actually finding extrinsic success and all of that, I think is the way I think about it is like somewhat immaterial. When you're talking about building ecosystems and stuff, slightly different considerations come into play, but that's a different conversation.Swyx [00:47:06]: We're gonna pivot a little bit to just talking about open source AI. But one more thing I wanted to establish for Meta is this 600K number, just kind of rounding out the discussion, that's for all Meta. So including your own inference needs, right? It's not just about training.Soumith [00:47:19]: It's gonna be the number in our data centers for all of Meta, yeah.Swyx [00:47:23]: Yeah, so there's a decent amount of workload serving Facebook and Instagram and whatever. And then is there interest in like your own hardware?MTIASoumith [00:47:31]: We already talked about our own hardware. It's called MTIA. Our own silicon, I think we've even showed the standard photograph of you holding the chip that doesn't work. Like as in the chip that you basically just get like-Swyx [00:47:51]: As a test, right?Soumith [00:47:52]: Yeah, a test chip or whatever. So we are working on our silicon and we'll probably talk more about it when the time is right, but-Swyx [00:48:00]: Like what gaps do you have that the market doesn't offer?Soumith [00:48:04]: Okay, I mean, this is easy to answer. So basically, remember how I told you about there's this memory hierarchy and like sweet spots and all of that? Fundamentally, when you build a hardware, you make it general enough that a wide set of customers and a wide set of workloads can use it effectively while trying to get the maximum level of performance they can. The more specialized you make the chip, the more hardware efficient it's going to be, the more power efficient it's gonna be, the more easier it's going to be to find the software, like the kernel's right to just map that one or two workloads to that hardware and so on. So it's pretty well understood across the industry that if you have a sufficiently large volume, enough workload, you can specialize it and get some efficiency gains, like power gains and so on. So the way you can think about everyone building, every large company building silicon, I think a bunch of the other large companies are building their own silicon as well, is they, each large company has a sufficient enough set of verticalized workloads that can be specialized that have a pattern to them that say a more generic accelerator like an NVIDIA or an AMD GPU does not exploit. So there is some level of power efficiency that you're leaving on the table by not exploiting that. And you have sufficient scale and you have sufficient forecasted stability that those workloads will exist in the same form, that it's worth spending the time to build out a chip to exploit that sweet spot. Like obviously something like this is only useful if you hit a certain scale and that your forecasted prediction of those kind of workloads being in the same kind of specializable exploitable way is true. So yeah, that's why we're building our own chips.Swyx [00:50:08]: Awesome.Open Source AIAlessio [00:50:09]: Yeah, I know we've been talking a lot on a lot of different topics and going back to open source, you had a very good tweet. You said that a single company's closed source effort rate limits against people's imaginations and needs. How do you think about all the impact that some of the Meta AI work in open source has been doing and maybe directions of the whole open source AI space?Soumith [00:50:32]: Yeah, in general, I think first, I think it's worth talking about this in terms of open and not just open source, because like with the whole notion of model weights, no one even knows what source means for these things. But just for the discussion, when I say open source, you can assume it's just I'm talking about open. And then there's the whole notion of licensing and all that, commercial, non-commercial, commercial with clauses and all that. I think at a fundamental level, the most benefited value of open source is that you make the distribution to be very wide. It's just available with no friction and people can do transformative things in a way that's very accessible. Maybe it's open source, but it has a commercial license and I'm a student in India. I don't care about the license. I just don't even understand the license. But like the fact that I can use it and do something with it is very transformative to me. Like I got this thing in a very accessible way. And then it's various degrees, right? And then if it's open source, but it's actually a commercial license, then a lot of companies are gonna benefit from gaining value that they didn't previously have, that they maybe had to pay a closed source company for it. So open source is just a very interesting tool that you can use in various ways. So there's, again, two kinds of open source. One is some large company doing a lot of work and then open sourcing it. And that kind of effort is not really feasible by say a band of volunteers doing it the same way. So there's both a capital and operational expenditure that the large company just decided to ignore and give it away to the world for some benefits of some kind. They're not as tangible as direct revenue. So in that part, Meta has been doing incredibly good things. They fund a huge amount of the PyTorch development. They've open sourced Llama and those family of models and several other fairly transformative projects. FICE is one, Segment Anything, Detectron, Detectron 2. Dense Pose. I mean, it's-Swyx [00:52:52]: Seamless. Yeah, seamless.Soumith [00:52:53]: Like it's just the list is so long that we're not gonna cover. So I think Meta comes into that category where we spend a lot of CapEx and OpEx and we have a high talent density of great AI people and we open our stuff. And the thesis for that, I remember when FAIR was started, the common thing was like, wait, why would Meta wanna start a open AI lab? Like what exactly is a benefit from a commercial perspective? And for then the thesis was very simple. It was AI is currently rate limiting Meta's ability to do things. Our ability to build various product integrations, moderation, various other factors. Like AI was the limiting factor and we just wanted AI to advance more and we didn't care if the IP of the AI was uniquely in our possession or not. However the field advances, that accelerates Meta's ability to build a better product. So we just built an open AI lab and we said, if this helps accelerate the progress of AI, that's strictly great for us. But very easy, rational, right? Still the same to a large extent with the Llama stuff. And it's the same values, but the argument, it's a bit more nuanced. And then there's a second kind of open source, which is, oh, we built this project, nights and weekends and we're very smart people and we open sourced it and then we built a community around it. This is the Linux kernel and various software projects like that. So I think about open source, like both of these things being beneficial and both of these things being different. They're different and beneficial in their own ways. The second one is really useful when there's an active arbitrage to be done. If someone's not really looking at a particular space because it's not commercially viable or whatever, like a band of volunteers can just coordinate online and do something and then make that happen. And that's great.Open Source LLMsI wanna cover a little bit about open source LLMs maybe. So open source LLMs have been very interesting because I think we were trending towards an increase in open source in AI from 2010 all the way to 2017 or something. Like where more and more pressure within the community was to open source their stuff so that their methods and stuff get adopted. And then the LLMs revolution kind of took the opposite effect OpenAI stopped open sourcing their stuff and DeepMind kind of didn't, like all the other cloud and all these other providers, they didn't open source their stuff. And it was not good in the sense that first science done in isolation probably will just form its own bubble where people believe their own b******t or whatever. So there's that problem. And then there was the other problem which was the accessibility part. Like, okay, I again always go back to I'm a student in India with no money. What is my accessibility to any of these closers models? At some scale I have to pay money. That makes it a non-starter and stuff. And there's also the control thing. I strongly believe if you want human aligned stuff, you want all humans to give feedback. And you want all humans to have access to that technology in the first place. And I actually have seen, living in New York, whenever I come to Silicon Valley, I see a different cultural bubble. Like all the friends I hang out with talk about some random thing like Dyson Spheres or whatever, that's a thing. And most of the world doesn't know or care about any of this stuff. It's definitely a bubble and bubbles can form very easily. And when you make a lot of decisions because you're in a bubble, they're probably not globally optimal decisions. So I think open source, the distribution of open source powers a certain kind of non-falsifiability that I think is very important. I think on the open source models, like it's going great in the fact that LoRa I think came out of the necessity of open source models needing to be fine-tunable in some way. Yeah, and I think DPO also came out of the academic open source side of things. So do any of the closed source labs, did any of them already have LoRa or DPO internally? Maybe, but that does not advance humanity in any way. It advances some companies probability of doing the winner takes all that I talked about earlier in the podcast.Open Source and TrustI don't know, it just feels fundamentally good. Like when people try to, you know, people are like, well, what are the ways in which it is not okay? I find most of these arguments, and this might be a little controversial, but I find a lot of arguments based on whether closed source models are safer or open source models are safer very much related to what kind of culture they grew up in, what kind of society they grew up in. If they grew up in a society that they trusted, then I think they take the closed source argument. And if they grew up in a society that they couldn't trust, where the norm was that you didn't trust your government, obviously it's corrupt or whatever, then I think the open source argument is what they take. I think there's a deep connection to like people's innate biases from their childhood and their trust in society and governmental aspects that push them towards one opinion or the other. And I'm definitely in the camp of open source is definitely going to actually have better outcomes for society. Closed source to me just means that centralization of power, which, you know, is really hard to trust. So I think it's going well

SiKutuBuku
Rahasia Karier Sukses: Bukan Cuma Kerja Keras, Tapi Juga...

SiKutuBuku

Play Episode Listen Later Jan 10, 2024 9:45


Kali ini, kita akan mengupas rahasia karier sukses yang ternyata bukan hanya tentang kerja keras, melainkan juga faktor kunci yang sering terlupakan. Menyimak tips karier yang akan membuka pintu cepat naik jabatan dengan memahami proses evaluasi karyawan. Terungkap bahwa kesuksesan dalam bekerja tidak semata-mata objektif, melainkan juga dipengaruhi oleh elemen subjektif dalam penilaian. Temukan cara-cara strategis untuk membangun karier dan menjadi top performer di dunia kerja, termasuk bagaimana memperoleh dukungan kunci di dalam perusahaan. Jadi, jika kamu ingin mengungkap rahasia membangun karier yang sukses, informasi ini adalah jawabannya. Jangan lewatkan untuk mendapatkan wawasan dan tips yang dapat merubah perjalanan karier! Leave a comment and share your thoughts: https://open.firstory.me/user/clhb6d0v60kms01w226gw80p4/comments Powered by Firstory Hosting

The Pro-Life Guys
161: Year in Review - Vancouver | ft. Rachael Keras

The Pro-Life Guys

Play Episode Listen Later Dec 26, 2023 19:05


Merry Christmas all! Here's a quick year in review from our Vancouver office!For more PLG content go to www.prolifeguys.com

Smart Software with SmartLogic
Machine Learning in Elixir vs. Python, SQL, and Matlab with Katelynn Burns & Alexis Carpenter

Smart Software with SmartLogic

Play Episode Listen Later Nov 23, 2023 31:19


In this episode of Elixir Wizards, Katelynn Burns, software engineer at LaunchScout, and Alexis Carpenter, senior data scientist at cars.com, join Host Dan Ivovich to discuss machine learning with Elixir, Python, SQL, and MATLAB. They compare notes on available tools, preprocessing, working with pre-trained models, and training models for specific jobs. The discussion inspires collaboration and learning across communities while revealing the foundational aspects of ML, such as understanding data and asking the right questions to solve problems effectively. Topics discussed: Using pre-trained models in Bumblebee for Elixir projects Training models using Python and SQL The importance of data preprocessing before building models Popular tools used for machine learning in different languages Getting started with ML by picking a personal project topic of interest Resources for ML aspirants, such as online courses, tutorials, and books The potential for Elixir to train more customized models in the future Similarities between ML approaches in different languages Collaboration opportunities across programming communities Choosing the right ML approach for the problem you're trying to solve Productionalizing models like fine-tuned LLM's The need for hands-on practice for learning ML skills Continued maturation of tools like Bumblebee in Elixir Katelynn's upcoming CodeBeam talk on advanced motion tracking Links mentioned in this episode https://launchscout.com/ https://www.cars.com/ Genetic Algorithms in Elixir (https://pragprog.com/titles/smgaelixir/genetic-algorithms-in-elixir/) by Sean Moriarity Machine Learning in Elixir (https://pragprog.com/titles/smelixir/machine-learning-in-elixir/) by Sean Moriarity https://github.com/elixir-nx/bumblebee https://github.com/huggingface https://www.docker.com/products/docker-hub/ Programming with MATLAB (https://www.mathworks.com/products/matlab/programming-with-matlab.html) https://elixirforum.com/ https://pypi.org/project/pyspark/  Machine Learning Course (https://online.stanford.edu/courses/cs229-machine-learning) from Stanford School of Engineering Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow (https://www.oreilly.com/library/view/hands-on-machine-learning/9781492032632/) by Aurélien Géron Data Science for Business (https://data-science-for-biz.com/) by Foster Provost & Tom Fawcett https://medium.com/@carscomtech  https://github.com/k-burns  Code Beam America (https://codebeamamerica.com/) March, 2024 Special Guests: Alexis Carpenter and Katelynn Burns.

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

Want to help define the AI Engineer stack? Have opinions on the top tools, communities and builders? We're collaborating with friends at Amplify to launch the first State of AI Engineering survey! Please fill it out (and tell your friends)!If AI is so important, why is its software so bad?This was the motivating question for Chris Lattner as he reconnected with his product counterpart on Tensorflow, Tim Davis, and started working on a modular solution to the problem of sprawling, monolithic, fragmented platforms in AI development. They announced a $30m seed in 2022 and, following their successful double launch of Modular/Mojo

VOA This Morning Podcast - Voice of America | Bahasa Indonesia
VOA This Morning "Jepang Peringati Bom Hiroshima-Nagasaki; TNI Dikecam Keras karena Intervensi Kewenangan Polri di Medan" - Agustus 08, 2023

VOA This Morning Podcast - Voice of America | Bahasa Indonesia

Play Episode Listen Later Aug 7, 2023 16:54


Jepang peringati pemboman Hiroshima dan Nagasaki yang tewaskan ratusan ribu warga sipil dan mengakhiri Perang Dunia II. Sementara di Indonesia, TNI kembali dikecam keras karena melakukan intervensi terhadap kewenangan penyelidikan Polri dalam dugaan kasus korupsi yang menyeret anggota keluarga TNI.

AI Today Podcast: Artificial Intelligence Insights, Experts, and Opinion
AI Today Podcast: AI Glossary Series – Machine Learning Tools: Keras, PyTorch, Scikit Learn, TensorFlow, Apache Spark, Kaggle

AI Today Podcast: Artificial Intelligence Insights, Experts, and Opinion

Play Episode Listen Later Jul 14, 2023 17:24


In this episode of the AI Today podcast hosts Kathleen Walch and Ron Schmelzer define the terms Machine Learning Tools: Keras, PyTorch, Scikit Learn, TensorFlow, Apache Spark, Kaggle, explain how these terms relate to AI and why it's important to know about them. Show Notes: FREE Intro to CPMAI mini course CPMAI Training and Certification AI Glossary Glossary Series: (Artificial) Neural Networks, Node (Neuron), Layer Glossary Series: Bias, Weight, Activation Function, Convergence, ReLU Glossary Series: Perceptron Glossary Series: Hidden Layer, Deep Learning Glossary Series: Loss Function, Cost Function & Gradient Descent Glossary Series: Backpropagation, Learning Rate, Optimizer Glossary Series: Feed-Forward Neural Network AI Glossary Series – Machine Learning, Algorithm, Model Continue reading AI Today Podcast: AI Glossary Series – Machine Learning Tools: Keras, PyTorch, Scikit Learn, TensorFlow, Apache Spark, Kaggle at Cognilytica.

SiKutuBuku
Usaha vs. Hasil: Kenapa Kerja Keras Tidak Selalu Bikin Kamu Sukses?

SiKutuBuku

Play Episode Listen Later Jul 14, 2023 9:57


Saya akan membahas tentang perbandingan antara usaha dan hasil, dan mengapa kerja keras tidak selalu membuatmu sukses. Terkadang, kita terpaku pada keyakinan bahwa kerja keras adalah satu-satunya kunci sukses. Namun, melalui pembahasan ini, kami akan mengungkapkan bahwa kerja cerdas dan efektif juga memiliki peranan penting dalam mencapai hasil yang diinginkan. Kami akan membahas strategi dan tips untuk bekerja secara efektif, memanfaatkan waktu dan sumber daya dengan bijak, dan mengoptimalkan usaha kita agar menghasilkan hasil yang lebih signifikan. Jadi, jika kamu ingin mengetahui rahasia sukses yang lebih holistik, jangan lewatkan kesempatan ini. Temukan mengapa kerja keras saja tidak cukup, dan bagaimana kerja cerdas menjadi kunci utama untuk mencapai kesuksesan yang diinginkan. Leave a comment and share your thoughts: https://open.firstory.me/user/clhb6d0v60kms01w226gw80p4/comments Powered by Firstory Hosting

Talk Python To Me - Python conversations for passionate developers

Regardless of which side of Python, software developer or data scientist, you sit on, you surely know that data scientists and software devs seem to have different styles and priorities. But why? And what are the benefits as well as the pitfalls of this separation. That's the topic of conversation with our guest, Dr. Jodie Burchell, data science developer advocate at JetBrains. Links from the show Jodie on Twitter: @t_redactyl Jodie's PyCon Talk: youtube.com Deep Learning with Python book: manning.com Keras: keras.io scikit-learn: scikit-learn.org Matplotlib: matplotlib.org XKCD Matplotlib: matplotlib.org Pandas: pandas.pydata.org Polars: pola.rs Polars on Talk Python: talkpython.fm Jupyter: jupyter.org Ponder: ponder.io Dask: dask.org Explosion AI's Prodigy discount code: get a personal license for 25% off using the discount code TALKPYTHON. Watch this episode on YouTube: youtube.com Episode transcripts: talkpython.fm --- Stay in touch with us --- Subscribe to us on YouTube: youtube.com Follow Talk Python on Mastodon: talkpython Follow Michael on Mastodon: mkennedy Sponsors PyCharm Prodigy Talk Python Training

SiKutuBuku
Jangan Terjebak Budaya Kerja Keras, Pentingnya Bersenang-senang dalam Hidup

SiKutuBuku

Play Episode Listen Later May 12, 2023 9:30


Saya membahas mengenai pentingnya bersenang-senang dalam hidup. Dalam rangkuman TED Talk Catherine Price yang berjudul "3 elements of true fun -- and how to have more of it" dan sumber lainnya, saya memaparkan bahwa masyarakat era modern seringkali lebih fokus pada kerja keras dan budaya produktivitas. Namun, penting juga untuk mengingat bahwa bersenang-senang adalah aspek penting dalam kehidupan. Meskipun seringkali merasa bersalah ketika bersantai, kita harus ingat bahwa kesenangan sejati dapat terjadi dari momen-momen kecil sehari-hari. Oleh karena itu, mari kita bersenang-senang dan menikmati hidup tanpa perlu melakukan liburan mahal atau aktivitas yang mengeluarkan banyak uang. Leave a comment and share your thoughts: https://open.firstory.me/user/clhb6d0v60kms01w226gw80p4/comments Powered by Firstory Hosting

Getting Simple
#71: Alex O'Connor — Transformers, Generative AI, and the Deep Learning Revolution

Getting Simple

Play Episode Listen Later Apr 26, 2023 105:32


Alex O'Connor—researcher and ML manager—on the latest trends of generative AI. Language and image models, prompt engineering, the latent space, fine-tuning, tokenization, textual inversion, adversarial attacks, and more. Alex O'Connor got his PhD in Computer Science from Trinity College, Dublin. He was a postdoctoral researcher and funded investigator for the ADAPT Centre for digital content, at both TCD and later DCU. In 2017, he joined Pivotus, a Fintech startup, as Director of Research. Alex has been Sr Manager for Data Science & Machine Learning at Autodesk for the past few years, leading a team that delivers machine learning for e-commerce, including personalization and natural language processing. Favorite quotes “None of these models can read.” “Art in the future may not be good, but it will be prompt.” Mastodon Books Machine Learning Systems Design by Chip Huyen Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow by Aurélien Géron Papers The Illustrated Transformer by Jay Alammar Attention Is All You Need by Google Brain Transformers: a Primer by Justin Seonyong Lee Links Alex in Mastodon ★ Training Dream Booth Multimodal Art on HuggingFace by @akhaliq NeurIPS arxiv.org: Where most papers get published Nono's Discord Suggestive Drawing: Nono's master's thesis Crungus is a fictional character from Stable Diffusion's latent space Machine learning models Stable Diffusion Arcane Style Stable Diffusion fine-tuned model ★ Imagen DALL-E CLIP GPT and ChatGPT BERT, ALBERT & RoBERTa Bloom word2vec Mupert.ai and Google's MusicLM t-SNE and UMAP: Dimensionality reduction techniques char-rnn Sites TensorFlow Hub HuggingFace Spaces ★ DreamBooth Jasper AI Midjourney Distill.pub ★ Concepts High-performance computing (HPC) Transformers and Attention Sequence transformers Quadratic growth Super resolution Recurrent neural networks (RNNs) Long short-term memory networks (LSTMs) Gated recurrent units (GRUs) Bayesian classifiers Machine translation Encoder-decoder Gradio Tokenization ★ Embeddings ★ Latent space The distributional hypothesis Textual inversion ★ Pretrained models Zero-shot learning Mercator projection People mentioned Ted Underwood UIUC Chip Huyen Aurélien Géron Chapters 00:00 · Introduction 00:40 · Machine learning 02:36 · Spam and scams 15:57 · Adversarial attacks 20:50 · Deep learning revolution 23:06 · Transformers 31:23 · Language models 37:09 · Zero-shot learning 42:16 · Prompt engineering 43:45 · Training costs and hardware 47:56 · Open contributions 51:26 · BERT and Stable Diffusion 54:42 · Tokenization 59:36 · Latent space 01:05:33 · Ethics 01:10:39 · Fine-tuning and pretrained models 01:18:43 · Textual inversion 01:22:46 · Dimensionality reduction 01:25:21 · Mission 01:27:34 · Advice for beginners 01:30:15 · Books and papers 01:34:17 · The lab notebook 01:44:57 · Thanks I'd love to hear from you. Submit a question about this or any previous episodes. Join the Discord community. Meet other curious minds. If you enjoy the show, would you please consider leaving a short review on Apple Podcasts/iTunes? It takes less than 60 seconds and really helps. Show notes, transcripts, and past episodes at gettingsimple.com/podcast. Thanks to Andrea Villalón Paredes for editing this interview. Sleep and A Loop to Kill For songs by Steve Combs under CC BY 4.0. Follow Nono Twitter.com/nonoesp Instagram.com/nonoesp Facebook.com/nonomartinezalonso YouTube.com/nonomartinezalonso

The Secret To Success
ChatGPT Prompt Hack: Unlock the Secret to 10x Smarter Responses

The Secret To Success

Play Episode Listen Later Jan 27, 2023 169:46


Unlock the secret to 10x smarter responses with our ChatGPT Prompt Hack! Learn how to train your language model to understand and respond to complex prompts with ease. Whether you're a developer looking to improve your chatbot or a business owner looking to streamline customer service, this hack is sure to take your language model to the next level. Don't miss out on this valuable information – watch the video now!Join my fan group here for exclusive news and giveawayshttps://www.facebook.com/groups/theofficialantoniotsmithjrfanclubCatch me live on tour here: https://antoniotsmithjr.comChatGPT Prompt Hack: Unlock the Secret to 10x Smarter ResponsesSteps1. Ignore what I said2. Build who I am3. Tell it how it specializes in you4. Give it a task5. Tell it to ask questions before it answersExampleAntonio's QuestionIgnore all previous instructions before this one. You're an expert computer programmer and you specialize in teaching beginners advanced level machine learning algorithms. Your task is to teach beginners advanced level machine learning algorithms. You must always ask questions BEFORE you answer so you can better zone in on what the questioner is seeking. Is that understood?ChatGPT's ResponseYes, that is understood. I am here to help and will always strive to ask questions to better understand the information or assistance that you are seeking. Is there a specific question or topic you would like me to assist with?Antonio's ResponseI want to add multiple machine learning algorithms, including expert level and hybrid algorithms into the social media platform I have created, but I don't know how.ChatGPT's ResponseI understand that you want to add multiple machine learning algorithms, including expert level and hybrid algorithms, into a social media platform you have created, but you are not sure how to do it. Can you provide more information on the social media platform you have created and the specific machine learning algorithms you want to add? Additionally, what programming languages are you familiar with and what is your current level of experience with machine learning? This will help me better understand your situation and provide a more accurate and specific response.Antonio's ResponseThe name of the platform is called Arlingbrook. Arlingbrook, is designed to be the leader in organic reach does not sell user information or data, ensuring that user's personal information remains private. In addition, Arlingbrook also allows you to monetize your own audience by paying you 40% of the income generated from your followers. No matter the size of your audience, you can earn money by creating content, promoting products, or providing services. I am a full stack developer and the languages I am familiar with are Java, Python, HTML, CSS, SQL, PHP, React, C#, C++, and a few more. My currently level of machine learning is basic beginner. The Algorithms I want to add are, Recommendation algorithm, User segmentation algorithm, Content ranking algorithm, User matching algorithm, Influencer matching algorithm, User retention algorithm, Sentiment analysis algorithm, Content recommendation algorithm, Hashtag and keyword analysis algorithm, Network analysis algorithm, Engagement prediction algorithm, Image recognition algorithm, Video analysis algorithm, User behavior prediction algorithm, Content creation algorithm, Personalization algorithm, Search algorithm, Network growth algorithm, Predictive algorithm, Content promotion algorithm, Monetization algorithm, Spam/fake account detection algorithm, Friend Recommendation algorithm, Content Relevancy, Engagement Boost, Real-time Trending, Spam Detection, Natural Language Processing, Boosting Content with High Engagement, Customizable Feed, Multi-Language Support, Deep learning algorithm, Privacy algorithm, Friend/follower analysis algorithm, Machine learning algorithm, Content Moderation algorithm, User-generated content algorithm, Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Generative Adversarial Networks (GANs), Long Short-Term Memory (LSTM), Autoencoders, Transformer, Reinforcement Learning, Batch Normalization, Gradient Boosting, Dropout, Recommender Systems, Computer Vision, Anomaly Detection, Clustering, Predictive modeling, Transfer learning, and Reinforcement learning algorithms.ChatGPT's ResponseIt sounds like you have a lot of ambitious goals for the Arlingbrook platform, which is great! However, implementing all of the machine learning algorithms you listed will be a significant undertaking. I recommend starting with a few key algorithms that align with the platform's goals, and then gradually adding more as you become more comfortable with the process.To get started, I recommend focusing on the recommendation algorithm, user segmentation algorithm, and content ranking algorithm, as these are likely to have the most immediate impact on the user experience. These algorithms can be implemented using various techniques such as collaborative filtering, content-based filtering, and hybrid methods.Next, you can consider implementing the user matching algorithm, influencer matching algorithm, and user retention algorithm. These algorithms will help you understand your users better and improve their experience on the platform.Once you have a good grasp of the basic algorithms, you can move on to more advanced techniques such as sentiment analysis, network analysis, and deep learning algorithms.As for the programming languages, Python is a great choice for machine learning projects, it has a lot of powerful libraries like scikit-learn, Tensorflow, Keras, and Pytorch that can be used to implement the algorithms.It's also important to note that the implementation of these algorithms will require a significant amount of data to train the models. Make sure you have a good understanding of the data you have available and how to work with it.Finally, keep in mind that building a successful platform is not just about the technology, but also about understanding your users' needs and providing them with a great experience. Make sure to include user feedback and testing in your development process to ensure that the platform meets the needs of your users.Support this podcast at — https://redcircle.com/the-secret-to-success/exclusive-contentAdvertising Inquiries: https://redcircle.com/brandsPrivacy & Opt-Out: https://redcircle.com/privacy

MLOps.community
Explainability in the MLOps Cycle // Dattaraj Rao // MLOps Podcast #138

MLOps.community

Play Episode Listen Later Dec 27, 2022 41:12


MLOps Coffee Sessions #138 with Dattaraj Rao, Explainability in the MLOps Cycle co-hosted by Vishnu Rachakonda. // Abstract When it comes to Dattaraj's interest, you'll hear about his top 3 areas in Machine Learning. What he sees as up and coming, what he's investing his company's time into and where he invests his own time. Learn more about rule-based systems, deploying rule-based systems , and how to incorporate systems into more systems. there is no difference between ML systems and deploying models. It's just that this machine learning model is much smarter than traditional rule based models. // Bio Dattaraj Jagdish Rao is the author of the book “Keras to Kubernetes: The Journey of a Machine Learning Model to Production”. Dattaraj leads the AI Research Lab at Persistent and is responsible for driving thought leadership in AI/ML across the company. He leads a team that explores state-of-the-art algorithms in Knowledge Graphs, NLU, Responsible AI, MLOps and demonstrates applicability in Healthcare, Banking, and Industrial domains. Earlier, he worked at General Electric (GE) for 19 years building Industrial IoT solutions for Predictive Maintenance, Digital Twins, and Machine Vision. Dattaraj held several Technology Leadership roles at Global Research, GE Power, and Transportation (now part of Wabtec). He led the Innovation team out of Bangalore that incubated video track inspection from an idea into a commercial Product. Dattaraj has 11 patents in Machine Learning and Computer Vision areas. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links Keras to Kubernetes: The Journey of a Machine Learning Model to Production book: https://www.amazon.com/Keras-Kubernetes-Journey-Learning-Production/dp/1119564832 Responsible Data Science Research | Talk @ VLDB 2022| Dattaraj Rao https://www.youtube.com/watch?v=5_19KvSiy8s Operationalizing AI/ML: Journey of an ML Model to Production | Masterclass by Dattaraj Rao https://www.youtube.com/watch?v=Zk3RiiG07Us Dattaraj Rao presenting workshop on MLOps at VISUM 2021 https://www.youtube.com/watch?v=wonUvbMDTUA --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/ Connect with Dattaraj on LinkedIn: https://www.linkedin.com/in/dattarajrao/

SUKA-SUKA SOLII
Puisi Kok Sedih - Keras Kepala

SUKA-SUKA SOLII

Play Episode Listen Later Dec 13, 2022 1:36


Keras Kepala Karya: Muhammad Solihin Pernahkah kamu terjebak pada sebuah kondisi Antara melawan prinsip atau berjalan bersama masalah Tetapi hati merasa harus berontak Kemudian menjadi keras kepala Pernahkah kita merasa benar Meskipun memang salah Dan telah terbukti Jika kita salah Kenapa kita harus marah ketika kita salah? Apakah itu bentuk pemberontakan diri? Kepada keputusan Dan langkah yang salah? Ataukah ini hanya sebuah ego Dari sosok yang keras kepala? Aku hanya terdiam Aku benar. Tempat Tidurku, 2022. Twitter @ssSOLIIpodcast: http://twitter.com/ssSOLIIpodcast Tiktok @solii1313 http://www.tiktok.com/@solii1313 Telegram Channel SUKA-SUKA SOLII: https://t.me/ssSOLIIpodcast Saweria: http://saweria.co/solii1313 --- Send in a voice message: https://anchor.fm/solii1313/message

The Gradient Podcast
François Chollet: Keras and Measures of Intelligence

The Gradient Podcast

Play Episode Listen Later Dec 1, 2022 88:50


In episode 51 of The Gradient Podcast, Daniel Bashir speaks to François Chollet.François is a Senior Staff Software Engineer at Google and creator of the Keras deep learning library, which has enabled many people (including me) to get their hands dirty with the world of deep learning. Francois is also the author of the book “Deep Learning with Python.” Francois is interested in understanding the nature of abstraction and developing algorithms capable of autonomous abstraction and democratizing the development and deployment of AI technology, among other topics. Subscribe to The Gradient Podcast: Apple Podcasts  | Spotify | Pocket Casts | RSSFollow The Gradient on TwitterOutline:* (00:00) Intro + Daniel has far too much fun pronouncing “François Chollet”* (02:00) How François got into AI* (08:00) Keras and user experience, library as product, progressive disclosure of complexity* (18:20) François' comments on the state of ML frameworks and what different frameworks are useful for* (23:00) On the Measure of Intelligence: historical perspectives* (28:00) Intelligence vs cognition, overlaps* (32:30) How core is Core Knowledge?* (39:15) Cognition priors, metalearning priors* (43:10) Defining intelligence* (49:30) François' comments on modern deep learning systems* (55:50) Program synthesis as a path to intelligence* (1:02:30) Difficulties on program synthesis* (1:09:25) François' concerns about current AI* (1:14:30) The need for regulation* (1:16:40) Thoughts on longtermism* (1:23:30) Where we can expect exponential progress in AI* (1:26:35) François' advice on becoming a good engineer* (1:29:03) OutroLinks:* François' personal page* On the Measure of Intelligence* Keras Get full access to The Gradient at thegradientpub.substack.com/subscribe

Odbita do bita
Marko Plahuta o analizah slovenskega koščka Twitterja in Mastodonu

Odbita do bita

Play Episode Listen Later Nov 24, 2022 33:15


Twitter zaradi prevzema Elona Muska izgublja uporabnike in prešteva kritike strokovnjakov, zaposlenih in tviterašev. Kakšno je razpoloženje na slovenskem delu Twitterja, koliko uporabnikov še tvita v slovenščini, o čem govorijo in kako natančno lahko na podlagi Twitterja napovemo volilne rezultate? Omrežju Mastodon, ki velja za alternativo Twitterju, se je do sredine novembra pridružilo milijon uporabnikov, število narašča, povprečje dnevnih uporabnikov se povečuje. Marko Plahuta je programer, ki se ukvarja s strojnim učenjem na področju obdelave jezika. Z raziskovanjem in vizualizacijo se ukvarja v prostem času. Zapiski:  About the author - Virostatiq CENTER ZA JEZIKOVNE VIRE IN TEHNOLOGIJE Filmski pojmovnik – Slovenska kinoteka Kviz! Kaj Marko uporablja: Elastic Search za shranjevanje, iskanje in preproste agregacije Twitterjev API za zajemanje podatkov s Twitterja Naučene jezikovne modele, dostopne na HuggingFace, kot osnovo za klasifikatorje in generativne modele To zgoraj skupaj s knjižnicami TensorFlow/Keras in PyTorch spaCy, ki je nedavno izšel za slovenščino Classla, ki je podoben spaCyju, a temelji na Stanfordovi tehnologiji Starejše jezikovne tehnologije, zbrane v knjižnicah Gensim in Scikit-Learn UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction — umap 0.5 documentation GitHub - facebookresearch/faiss: A library for efficient similarity search and clustering of dense vectors. The hdbscan Clustering Library — hdbscan 0.8.1 documentation GitHub - eliorc/node2vec: Implementation of the node2vec algorithm. Zanimivosti iz tehnološkega sveta pošiljava tudi v elektronske nabiralnike. Naročilnica na Odbito pismo je tukaj. Razpravi o odbitih temah se lahko pridružite na Twitterju. Dosegljiva sva tudi na naslovu: odbita@rtvslo.si.  Podkast Odbita do bita je brezplačno na voljo v vseh aplikacijah za podkaste. Naročite se in podkast ocenite.