POPULARITY
In this special live episode, with guests Lisa Christensen, Hillary Miller and Christopher Lind, we explore their experience and deep expertise on the topics of L&D strategy and team structures - and it’s a masterclass of a conversation. Register for L&D Next 3rd - 6th March for free today at https://360learning.com/l-and-d-next/2025/ KEY TAKEAWAYS Understand what L&D is there to achieve. Team structures have to evolve and be flexible. L&D structure has to be right for achieving L&D aims, so it may be different from other teams in the organisation. Build adaptable teams by focusing on skillsets. Leverage data and hone in on one KPI. Understand which relationships and functions you need to focus on. That will evolve. Work holistically with other areas of the business. BEST MOMENTS “I've never seen a truly centralized model, ever.” “We can get hung up on the hierarchy of things and miss out.” “You gotta know who your players are, their bench strengths.” “We need better data capabilities in learning.” “Figure out what they care about and then lean on that.” Lisa Christensen Lisa leads McKinsey & Company's Learning Design and Development Center of Excellence, a global team of design experts, designing and building the learning that develops McKinsey Partners and professionals, enabling them to deliver incredible client impact. Lisa founded and leads McKinsey's Learning Research and Innovation Lab and sits on the global learning leadership team. https://www.linkedin.com/in/lisachristensen Christopher Lind Christopher Lind is a dynamic leader at the intersection of business, technology, and human experience, serving as an executive advisor, AI ethicist and sought-after voice in the L&D space. As a former Chief Learning Officer for ChenMed and GE Healthcare, Christopher has led transformative learning strategies that enhance workforce capability and business performance. A prominent commentator, speaker, and thought leader, he is known for his forward-thinking approach to digital learning, AI, and the evolving role of technology in talent development. https://www.linkedin.com/in/christopherlind Future Focused: https://christopherlind.substack.com Hillary Miller Hillary Miller is a seasoned Learning & Development leader currently heading L&D at HCA Healthcare. With a passion for driving workforce capability and business impact, she brings extensive experience in healthcare education and leadership development. Previously, as Chief Learning Officer at Penn State Health, Hillary led enterprise-wide learning strategies, fostering a culture of continuous development and innovation. https://www.linkedin.com/in/hillarybmiller VALUABLE RESOURCES https://podcasts.apple.com/gb/podcast/the-learning-development-podcast/id1466927523 L&D Master Class Series: https://360learning.com/blog/l-and-d-masterclass-home THE HOST David James David has been a People Development professional for more than 20 years, most notably as Director of Talent, Learning & OD for The Walt Disney Company across Europe, the Middle East & Africa. As well as being the Chief Learning Officer at 360Learning, David is a prominent writer and speaker on topics around modern and digital L&D. CONTACT METHOD Twitter: https://twitter.com/davidinlearning LinkedIn: https://www.linkedin.com/in/davidjameslinkedin L&D Collective: https://360learning.com/the-l-and-d-collective Blog: https://360learning.com/blog L&D Master Class Series: https://360learning.com/blog/l-and-d-masterclass-home
BIO: James “Jimmy” Milliron is Co-Founder & President of National Brokerage Atlantic, specializing in Wealth Enhancement, Estate Planning, and Asset Protection.STORY: Jimmy wanted to invest $100,000 in Bitcoin, but when he couldn't find an easy way to do it, he bought a car instead.LEARNING: Research and learn all you can about investment opportunities before investing. “Don't be afraid to pick up the phone and make a few calls. There's nothing like picking up the phone and talking to a real person on the other end instead of just texting them.”Jimmy Milliron Guest profileJames “Jimmy” Milliron is Co-Founder & President of National Brokerage Atlantic, specializing in Wealth Enhancement, Estate Planning, and Asset Protection. An insurance veteran, he previously served as Executive Vice President at NexTier Bank, building a $400 million premium finance portfolio. He holds a BA from VMI and various securities and insurance licenses.Worst investment everJimmy's worst investment is a mix between marrying a second wife and buying a car in 2016. He invested many resources in his second marriage, but it did not last that long.When Jimmy married his second ex-wife, he wanted to invest about $100,000 in Bitcoin. But he was busy and did not have time to research and learn more about Bitcoin. When Jimmy could not find an easy way to do it, he purchased a car instead with that cash.Lessons learnedGo the extra mile in research and learning about investment opportunities before investing.Consider all the investment options available.Actionable adviceIf you're young, seek advice from a mentor or your parents about what they would do instead of arbitrarily investing in a make-me-feel-good investment. Their guidance can be invaluable in navigating the complex world of investments.Jimmy's recommendationsJimmy recommends reading Donald Trump's Art of the Deal as a valuable resource for negotiation and decision-making.No.1 goal for the next 12 monthsJimmy's number one goal for the next 12 months is losing weight.Parting words “Thank you very much. Andrew and I wish everyone well.”Jimmy Milliron [spp-transcript] Connect with Jimmy MillironLinkedInWebsiteAndrew's booksHow to Start Building Your Wealth Investing in the Stock MarketMy Worst Investment Ever9 Valuation Mistakes and How to Avoid ThemTransform Your Business with Dr.Deming's 14 PointsAndrew's online programsValuation Master Class
How can technology reshape education to cater to every student's unique learning journey? Join us as we explore this question with Dr. Kristen DiCerbo, the Chief Learning Officer at Khan Academy, who shares her transformative insights into personalized and mastery-based learning. Discover how Khan Academy's new AI tool, "Khanmigo" is revolutionizing learning as it helps educators address the diverse needs of students worldwide, breaking free from the constraints of traditional educational models.Dr. DiCerbo highlights the challenges teachers face, such as managing large classrooms and addressing special education shortages, and how technology can alleviate these burdens. By leveraging platforms like Khan Academy, educators can provide tailored learning experiences that empower students and nurture creativity. We delve into the ways technology can reduce teacher burnout, foster stronger student-teacher connections, and transform teaching roles to be more focused on creating meaningful educational experiences and maintaining a healthy work-life balance.More About Our GuestDr. Kristen DiCerbo is the Chief Learning Officer at Khan Academy, where she leads the content, design, product management, and community support teams. Dr. DiCerbo's career has focused on embedding insights from education research into digital learning experiences. Prior to her role at Khan Academy, she was Vice-President of Learning Research and Design at Pearson, served as a research scientist supporting the Cisco Networking Academies, and worked as a school psychologist. Kristen has a Ph.D. in Educational Psychology from Arizona State University.Connect with Khan Academykhanmigo.aikhanacademy.orgGot a story to share or question you want us to answer? Send us a message!About the podcast:The KindlED Podcast explores the science of nurturing children's potential and creating empowering learning environments.Powered by Prenda Microschools, each episode offers actionable insights to help you ignite your child's love of learning. We'll dive into evidence-based tools and techniques that kindle young learners' curiosity, motivation, and well-being. Got a burning question?We're all ears! If you have a question or topic you'd love our hosts to tackle, please send it to podcast@prenda.com. Let's dive into the conversation together!Important links:• Connect with us on social • Subscribe to The Sunday Spark• Get our free literacy curriculum Interested in starting a microschool?Prenda provides all the tools and support you need to start and run an amazing microschool. Create a free Prenda World account to start designing your future microschool today. More info at ➡️ Prenda.com or if you're ready to get going ➡️ Start My Microschool
L&D leaders in large organizations often cite a lack of time as the main barrier to upskilling and reskilling. While increasing learning hours is important, the CLO Lift team found that time is not the most critical factor. Instead, the quality of learning has equal or greater influence on performance than the quantity of time spent. Rather than focusing solely on hours, business leaders and L&D professionals should emphasize value. This episode explores these insights with Lisa Christensen and Huw Newton-Hill. KEY TAKEAWAYS CLO Lift is a community of 20 CLOs who have got together to solve the biggest unresolved challenges in the L&D industry. You have to create true value before enough time will be allocated to employee learning. If you are not 100% in step with the business the training that you provide will be out of date and not relevant. The learning you deliver must genuinely enable people to make progress in their careers. L&D needs to stop only being order takers. Work more closely with managers to understand business challenges and be actively involved in finding solutions. Certain elements of people´s roles impact the way they behave. If a learning intervention cannot change that element speak up and ask those who can solve that underlying issue to do so. Create a learning culture, e.g. include skill acquisition in personal reviews. Democratize access to learning. Use generative AI to make the training more relevant in every role in every geography. Measure progress against actual business objectives. Start now. Create some small experiments and generate wins. You only need one win to create credibility and act as a springboard. BEST MOMENTS “Ensure that the whole body of L&D is in lockstep with the business.” “The more courage you lean into, the more credibility you build.” “Be bold and go forth.” “Any good action today is better than perfect action a month from now.” CLO Lift, From Time to Value report: https://learningforum.substack.com/p/clo-lift-from-time-to-value?r=ivpaq&utm_campaign=post&utm_medium=web&triedRedirect=true Lisa Christensen Bio: As Director of Learning Design and Innovation at McKinsey & Company, Lisa leads a global team focused on cutting-edge learning solutions. She founded McKinsey's Learning Research and Innovation Lab and is a recognized thought leader in learning. Lisa is a founding member of CLO Lift and was previously a senior leader at a learning design firm. Huw Newton-Hill Bio: Huw leads Attensi's US office, delivering AI-powered training solutions. A former strategy consultant at BCG, he now drives growth and innovation in L&D, contributing to forums like CLO Lift. VALUABLE RESOURCES The Learning And Development Podcast - https://podcasts.apple.com/gb/podcast/the-learning-development-podcast/id1466927523 L&D Master Class Series: https://360learning.com/blog/l-and-d-masterclass-home/ ABOUT THE HOST David James David has been a People Development professional for more than 20 years, most notably as Director of Talent, Learning & OD for The Walt Disney Company across Europe, the Middle East & Africa. As well as being the Chief Learning Officer at 360Learning, David is a prominent writer and speaker on topics around modern and digital L&D. CONTACT METHOD Twitter: https://twitter.com/davidinlearning/ LinkedIn: https://www.linkedin.com/in/davidjameslinkedin/ L&D Collective: https://360learning.com/the-l-and-d-collective/ Blog: https://360learning.com/blog/ L&D Master Class Series: https://360learning.com/blog/l-and-d-masterclass-home/
Get to know the person behind the research in this thoughtful conversation with acclaimed education researcher John Hattie. Join host Peter DeWitt and cohost Mike Nelson as they chat with John Hattie about his journey as a lifelong learner, how his Visible Learning research has resonated globally, and his goal for every educator to have a theory of learning. And, learn about John Hattie's highlight of his week--the days he spends with his grandson. Tune into this season seven finale of the Leaders Coaching Leaders Podcast for insights from an award-winning expert on enhancing student learning and achievement.Let us know what you think!
On this episode of the Getting Smart Podcast Mason is joined by repeat guest Aaron Schorn of Unrulr and two new guests for the Getting Smart Podcast, Kwaku Aning, Director of the Center of Innovation and Entrepreneurial Thinking at the San Diego Jewish Academy and Mike Yates, The Reinvention Lab at Teach for America. Links: Kwaku Aning LinkedIn Aaron Schorn LinkedIn Mike Yates LinkedIn San Diego Jewish Academy Reinvention Lab at Teach for America Designing Your Life MSCHF Hallcraft School Studio Unrulr Shoelace Learning Julia Dexter Changing the Subject d.school Scott Center for Entrepreneurship Bill Summers - Cañon City High School Capstone Consortium Big Picture Learning EyeCandy Previous episode with Aaron Schorn Lauryn Hill - Sister Act 2 Performance
Today on the show, Boston Children's Museum President and CEO Carole Charnow interviews Dr. Mitchel Resnick in the next installment of our Creativity Series. Mitch Resnick is the LEGO Papert Professor of Learning Research at the MIT Media Lab and develops new technologies and activities to engage people (particularly children) in creative learning experiences. His Lifelong Kindergarten research group developed the Scratch programming software and online community, used by millions of young people around the world. Carole talks with Mitch about his new project OctoStudio, the value of Kindergarten style learning, the 4 P's of Creative Learning, the relationship between technology and creative learning and more.
In the context of a cost-of-living crisis and increased child poverty, this podcast hears about the growing use of food banks, how they operate and the impact this has on children whose families use them. Full show notes and links: https://www.ucl.ac.uk/ioe/news/2023/dec/what-impact-do-food-banks-schools-have-childrens-learning-rftrw-s21e02
Jason Yosinski was a founding member of Uber AI Labs. He is also a co-founder of WinscapeAI a company dedicated to using custom sensor networks and machine learning to increase the efficiency and sustainability of wind farms. Jason holds a PhD in computer science from Cornell University. We talked about his experience at Uber AI, his research in deep learning, and ML for wind farms. Subscribe to Daliana's newsletter on www.dalianaliu.com for more on data science and career. Jason's Website: https://yosinski.com/ Jason's LinkedIn: https://www.linkedin.com/in/jasonyosinski/ Daliana's Twitter: https://twitter.com/DalianaLiu Daliana's LinkedIn: https://www.linkedin.com/in/dalianaliu (00:00:00) Introduction (00:06:06) His advice for Uber ML teams (00:16:03) From research to industry (00:20:24) ML for wind farms (00:25:40) Metrics for wind energy prediction (00:29:23) Start with a small dataset (00:32:00) ML in academia vs. the industry (00:33:24) Do you need a PhD for ML? (00:38:14) Daliana's story about grad school (00:41:37) The value of a PhD (00:43:13) ML Collective (00:48:36) Technical communication (00:57:21) ML Skillsets (00:59:45) Future of machine learning (01:05:23) Personal development: Hoffman process (01:15:13) Do things that excites you
Speaker: Andrew Burns, Vice President at BTS Andrew Burns is a Vice President at BTS responsible for partnering with clients to accelerate strategy alignment and culture change initiatives. His work spans the employee lifecycle, from developing assessment centers for hiring and leader development to defining what great leadership looks like within an organization's unique context. Andrew has extensive experience working in industries ranging from Aerospace to Oil and Gas, Utilities, and Software. Host: Rachel Cooke, Brandon Hall Group™ Rachel Cooke is Brandon Hall Group's Chief Operating Officer and Principal HCM Analyst. She is responsible for business operations, including client and member advisory services, marketing design, annual awards programs, conferences and the company's project management functions. She also leads Advancing Women in the Workplace and Diversity, Equity and Inclusion initiatives, research and events. Rachel worked in the HCM research industry for 20+ years and held several key management and executive positions within the Talent and Learning Research, and Performance Improvement industries.
To the Classroom: Conversations with Researchers & Educators
Today my guest is Dr. Margaret McKeown. We'll start our conversation discussing vocabulary development and explicit vocabulary instruction, including how to choose words for instruction, how to teach words so students understand them deeply, and how to help students build connections between words. Our conversation then shifts to the Questioning the Author instructional intervention, which focuses on developing comprehension through conversation and can be used to foster independence and discussion amongst students. Later, I'm joined by my colleagues Gina Dignon and Rosie Maurantonio for a conversation about how we'll bring what we learned to the classroom. ****Read a full transcript of this episode and learn more about the show and Jennifer Serravallo at JenniferSerravallo.comBringing Words to Life: Robust Vocabulary Instruction ****More about Dr. Margaret McKeown:Margaret G. McKeown, PhD, is Clinical Professor Emerita of Education at the University of Pittsburgh. Before her retirement, she was also a Senior Scientist at the University's Learning Research and Development Center. Her work addresses practical, current problems that classroom teachers and their students face. She has conducted research in the areas of learning, instruction, and teacher professional development in reading comprehension and vocabulary. Dr. McKeown is a recipient of the Outstanding Dissertation Award from the International Literacy Association, is a Fellow of the American Educational Research Association, and was inducted into the Reading Hall of Fame. She is coauthor of books including Bringing Words to Life, Second Edition; Creating Robust Vocabulary; Robust Comprehension Instruction with Questioning the Author; and Vocabulary Assessment to Support Instruction.Special thanks to Alex Van Rose for audio editing this episode. Support this showSupport the show
Advitya Gemawat ( https://www.microsoft.com/en-us/research/people/agemawat/ ) is a Machine Learning Research Engineer who currently works in the Responsible AI (RAI) team as part of the Azure Machine Learning (AML) and Azure AI Studio product lines, in the Azure AI Platform org. Prior to the RAI team, he was part of the 4th cohort of the Microsoft AI Development Acceleration Program (MAIDAP), where he worked with Azure Quality, Microsoft's Gray Systems Lab (GSL), Azure Edge & Platform, and Azure Machine Learning (AML). His research experience is at the intersection of Machine Learning & Scalable Systems. His work with GSL was identified as the Microsoft Global Hackathon Executive Challenge 2022 Winner and recipient of the Best Demonstration Award at VLDB 2022. His work with the AML team on expanding the RAI Dashboard to support Object Detection models was released in Public Preview at Microsoft Build 2023. Prior to Microsoft, Advitya graduated from UC San Diego with a Data Science major, where he contributed to Project Cerebro (a Layered Data Platform for scalable Deep Learning) and was advised by Professor Arun Kumar. There, his research won the 2021 ACM SIGMOD Student Research Abstract Competition and the Best Project Award as part of UCSD's Halıcıoğlu Data Science Institute (HDSI) Undergraduate Scholarship Program HDSI Scholarship Program. As part of an internship with VMware, his work on Massively Parallel Automated Model Building for Deep Learning was included in Apache MADlib 1.18.0 release. Opinions expressed in the video solely reflect Advitya's views and not the views of his employer. Follow Advitya on the handles below: LinkedIn: linkedin.com/in/agemawat YouTube: youtube.com/@AdvityaGemawat Microsoft Research: microsoft.com/research/people/agemawat Support the show
Peter's Film Course - https://thinkinginenglish.as.me/peters-film-course Thinking in English Classes Website - https://thinkinginenglish.link/ --------- What insights can science provide about language learning? And how can we use this research to learn English? Keep listening to find out! TRANSCRIPT - https://thinkinginenglish.blog/2023/07/19/252-science-and-english-learning-research-backed-tips-and-tricks-to-help-you-study/ ----------- My Links 7 Day FREE CONVERSATION CLUB TRIAL - https://www.patreon.com/thinkinginenglish JOIN THE CONVERSATION CLUB -- https://www.patreon.com/thinkinginenglish Take a Class (Use code TRIAL50 for 50% off) - https://thinkinginenglish.link/ Buy Me a Coffee - https://www.buymeacoffee.com/dashboard NEW YOUTUBE Channel!!! - https://www.youtube.com/@thinkinginenglishpodcast INSTAGRAM - thinkinginenglishpodcast (https://www.instagram.com/thinkinginenglishpodcast/) Blog - thinkinginenglish.blog -------- Vocabulary To distribute (v) - to spread out tasks, resources, or information over intervals of time, space, or people. Optimal (adj) - the most favourable or best possible. Retention (n) - the ability to retain or remember information. Recall (n) - the ability to remember things. Input (n) - information or data that is fed into a system or the brain. Memorisation (n) - the act or process of learning something so that you will remember it exactly: Literacy (n) - the ability to read and write. Immediate feedback (n) - prompt information given in response to an action or performance. --- Send in a voice message: https://podcasters.spotify.com/pod/show/thinking-english/message Support this podcast: https://podcasters.spotify.com/pod/show/thinking-english/support
Omar Flores (@OmarUFlorez en Twitter) es un experto en Modelos de Lenguaje, empleando señales sociales para optimizar la interpretación y expresión semántica de estos modelos. Sus métodos han fortalecido notablemente tareas como la búsqueda y recomendación. Ha trabajado como Machine Learning Research Scientist en Twitter, Senior Research Manager en Capital One y Research Scientist en Intel. Recibió el Premio a la Innovación de IBM Research en Análisis de Datos Escalables que financió su tesis doctoral. Este episodio profundiza en los avances y retos de la inteligencia artificial (IA), desde el modelo maestro, hasta los progresos en los modelos de lenguaje y la gestión de memoria. Se discute sobre la integración social de la IA y la capacidad de los modelos generativos para resolver problemas. Se aborda sobre memoria in context, la estrategia de IA de Perú, la diversidad de datos y las implicaciones éticas.
feedback @ ryan@soulsearching.in EPISODE LINKS: Website : https://sebastianraschka.com/ Linkedin : https://www.linkedin.com/in/sebastianraschkaTwitter : https://twitter.com/rasbtGitHub : https://github.com/rasbtYouTube : https://www.youtube.com/c/sebastianraschka PODCAST INFO: Podcast website: https://anchor.fm/ryandsouza Apple Podcasts: https://apple.co/3NQhg6S Spotify: https://spoti.fi/3qJ3tWJ Amazon Music: https://amzn.to/3P66j2B Google Podcasts: https://bit.ly/3am7rQc Gaana: https://bit.ly/3ANS4v1 RSS: https://anchor.fm/s/609210d4/podcast/rss
Many school librarians are exploring how they can use ChatGPT and generative AI to support teaching, learning, research, and creativity. We invited three librarians to share their early discoveries and some innovative ways they are using this rapidly expanding technology. Follow on Twitter: @joycevalenza @aelissmalespina@lucasjmaxwell @sgthomas1973 @bamradionetwork @jonHarper70bd Joyce Valenza is Associate Professor, at Rutgers University, SC&I and wrote the NeverendingSearch Blog for School Library Journal (now on hiatus), and contributes to several other library and tech publications. She speaks globally about the thoughtful use of technology in learning and the power of librarians to lead. Joyce was honored with the American Association of School Librarians' Distinguished Service Award and named an AASL Social Media Leadership Luminary. She is a Milken Educator and an American Memory Fellow. Joyce earned her doctorate in information science from the University of North Texas. Lucas Maxwell has been working with youth in libraries for fifteen years. He currently works as a school librarian in South London, UK. In 2017 he was named the UK's School Librarian of the Year, and in 2022, he was called the UK's Reading for Pleasure Champion by the UK's Literacy Association. Elissa Malespina is the librarian at Union High School in Union, NJ, and serves as a Board of Education Member for the South Orange - Maplewood School District, where she has been instrumental in helping to pass policies that strengthen the district's commitment to have diverse resources and materials available to all students. Elissa is at the forefront of using Web 2.0 resources and tools like Augmented Reality to make the experience in my library more interactive. Her work has been featured in NPR, School Library Journal, Publishers Weekly, and the PBS Documentary Film School Sleuth: The Case of The Wired Classroom. *The views expressed by Malespina do not represent her employer.
Are you interested in radical collaboration as an education form? Our summary today works with the chapter titled Radical collaboration: flipping the paradigm on learning from 2021 by Kelly Boucher, from the book titled Collaboration, visionaries share a new way of living. This episode is a great preparation for our next interviewee, Kelly Boucher in episode 114. Since we are investigating the future of cities, I thought it would be interesting to see how to think otherwise of life and place. This chapter presents radically collaborative ways of living well with the world and others in glorious multiplicity and complexity. As the most important things, I would like to highlight 3 aspects: Pedagogy is a term usually used to describe the art of teaching and learning or the agency that joins teaching and learning, but it can evolve to an active dialogue and being in relation with the world. Being in relation can mean, according to Aboriginals, that the boundaries between humans and nature are blurred because everything is animate. Through being in relation and radical re-learning, we take up our responsibility and step into radical collaboration with the world. You can find the book through this link. Connecting episodes you might be interested in: No.099 - Interview with Noel Tighe about leaving ego outside for the best outcomes in design processes; No.114 - Interview with Kelly Boucher about radical collaboration and unlearning; You can find the transcript through this link. What wast the most interesting part for you? What questions did arise for you? Let me know on Twitter @WTF4Cities or on the wtf4cities.com website where the shownotes are also available. I hope this was an interesting episode for you and thanks for tuning in. Music by Lesfm from Pixabay
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Touch reality as soon as possible (when doing machine learning research), published by Lawrence Chan on January 3, 2023 on The AI Alignment Forum. Related to: Making Beliefs Pay Rent, The Feeling of Idea Scarcity, Micro-Feedback Loops and Learning, The Three Stages of Rigor, Research as a Stochastic Decision Process, Chapter 22 of HPMOR. TL;DR: I think new machine learning researchers often make one of two kinds of mistakes: not making enough contact with reality, and being too reluctant to form gears-level models of ML phenomena. Stereotypically, LW/AF researchers tend to make the former mistake, while academic and industry researchers tend to make the latter kind. In this post, I discuss what I mean by “touching reality” and why it's important, speculate a bit on why people don't do this, and then give concrete suggestions. Epistemic status: I'm pretty frustrated with how slow I write, so this is an experiment in writing fast as opposed to carefully. That being said, this is ~the prevailing wisdom amongst many ML practitioners and academics, and similar ideas have been previously discussed in the LessWrong/Alignment Forum communities, so I'm pretty confident that it's directionally correct. I also believe (less confidently) that this is good advice for most kinds of research or maybe even for life in general. Acknowledgments: Thanks to Adrià Garriga-Alonso for feedback on a draft of this post and Justis Mills for copyediting help. Introduction: Broadly speaking, I think new researchers in machine learning tend to make two kinds of mistakes: Not making contact with reality. This is the failure mode where a new researcher reads a few papers that their friends are excited about, forms an ambitious hypothesis about how to solve a big problem in machine learning, and then spends months drafting a detailed plan. Unfortunately, after months of effort, our new researcher realizes that the components they were planning to use do not work nearly as well as expected, and as a result they've wasted months of effort on a project that wasn't going to succeed. Not being willing to make gears-level models. This is the failure mode where a new researcher decides to become agnostic to why anything happens, and believes empirical results and only empirical results even when said results don't “make sense” on reflection. The issue here is that they tend to be stuck implementing an inefficient variant of grad student descent, only able to make small amounts of incremental progress via approximate blind search, and end up doing whatever is popular at the moment. That's not to say that these mistakes are mutually exclusive: embarrassingly, I think I've managed to fail in both ways in the past. That being said, this post is about the first failure mode, which I think is far more common in our community than the second. (Though I might write about the second if there's enough interest!) Here, by “touching reality”, I mean running experiments where you check that your beliefs are right, either via writing code and running empirical ML experiments, or (less commonly) grounding your ideas either in a detailed formalism (to the level where you can write proofs for new, non-trivial theorems about said ideas). I don't think writing code or inventing a formalism qualify by themselves (though they are helpful); touching reality requires receiving actual concrete feedback on your ideas. Why touch reality? I think there's four main reasons why you should do this: Your ideas may be bad When you're new to a field, it's probably the case that you don't fully understand all of the key results and concepts in the field. As a result, it's very likely the case that the ideas you come up with are bad. This is especially true for fields like machine learning that have significant amounts of tacit knowledge. ...
So it's been a couple of weeks since I've released a podcast episode, but in case you didn't know, all of the episodes are streamed on Crowdcast, and I also post them to YouTube. So the last few episodes were visual; last week's episode was based on icebreakers using slides, and before that, I streamed live from DevLearn in Las Vegas. I can't format those appropriately for a podcast, but you can find them on www.TheTLDC.com or on our YouTube channel. This episode was great for a podcast. Jo Cook did a TLDC takeover to announce her and Jane Daly releasing a research report on Hybrid and Virtual Learning. Jo is an amazing Virtual Learning facilitator, and the report findings led to lots of community discussion, which you'll hear Jo share. Give it a listen - there's lots to learn here for everyone from VILT Producers to Zoom attendees. You can download the report here: https://virtualresearchinsights.com/report2022/ And check out their infographic for the report here: https://virtualresearchinsights.com/2022/09/26/infographic-for-improving-virtual-and-hybrid-learning/
A new diversity toolkit is providing resources for long-term care organizations to help raise awareness of issues of equity, diversity and inclusion. The Embracing Diversity: A Toolkit for Supporting Inclusion in Long-Term Care Homes is an interactive resource with print and online components that gives long-term care homes practical steps to nurture diverse and welcoming communities.We spoke with Michelle Fleming and Ashley Flanagan about the toolkit and the importance of understanding issues of equity, diversity and inclusion in long-term care. Michelle Fleming is a senior knowledge broker at the Ontario Centres for Learning Research and Innovation in Long-Term Care. Dr. Ashley Flanagan is a research fellow in Diversity in Aging at the National Institute on Aging.For more information about the toolkit, contact Michelle at: MFleming@bruyere.orgLearn more about the Strengthening a Palliative Approach to Long-Term Care project at: https://spaltc.ca/
Jak wygląda praca Machine Learning Research Engineera w Allegro? Jak udaje nam się efektywnie operować na ogromnej ilości danych oraz milionach parametrów? Czym są modele językowe trenowane na korpusie danych pochodzących z Allegro i jaki mają wpływ na realizowane projekty?Jak współpracujemy ze środowiskiem naukowym? I wreszcie - gdzie i w jakiej formie można znaleźć wiedzę, którą dzieli się nasz zespół Machine Learning Research?O tym wszystkim rozmawialiśmy z kolejnym gościem Allegro Tech Podcast, czyli Riccardo Belluzzo, który pracuje jako Machine Learning Research Engineer w Allegro. CONTACT LINKS / LINKS MENTIONED IN THE PODCAST:https://www.linkedin.com/in/riccardo-belluzzo/Allegro ML Research - https://ml.allegro.tech/Do you speak Allegro?” Large Scale Language Modeling - https://www.youtube.com/watch?v=6T-R4kgIbBsHerBERT na GitHub - https://github.com/allegro/HerBERT BIO:Riccardo is a Research Engineer in the Allegro ML team and specializes in Natural Language Processing and Understanding. Riccardo is also a music freak, playing guitar in his free time and running a podcast about underground music and emerging artists.
In this episode, we speak to Shawna Young and Mitchel Resnick of the Scratch Foundation, which runs the largest creative computing community in the world around the Scratch programming language. Recommended Resources:Mitch Resnick and Ken Robinson, Lifelong Kindergarten2021 Scratch Foundation Annual ReportShawna Young is the Executive Director of the Scratch Foundation. Before coming to Scratch, Young led the Duke University Talent Identification Program (Duke TIP), one of the largest academic talent searches, with over 450,000 K-12 students and over 3 million alumni. She also spearheaded the expansion of the Office of Engineering Outreach Programs (OEOP) at MIT, serving as the Executive Director for eight years. The OEOP provides rigorous educational opportunities in science, technology, engineering, and mathematics (STEM) to K-12 students from primarily underrepresented and underserved backgrounds. Young started her career as a public high school science teacher in North Carolina, then working as a curriculum developer at the Educational Development Center.Mitchel Resnick is the LEGO Papert Professor of Learning Research and Director of the Lifelong Kindergarten group at the Massachusetts Institute of Technology (MIT) Media Lab, which developed the Scratch programming software and online community, the world's leading coding platform for kids. His group has also collaborated for many years with the LEGO Company and the LEGO Foundation on the development of new educational ideas and products, including LEGO Mindstorms and LEGO WeDo robotics kits. Resnick co-founded the Computer Clubhouse project, an international network of 100 after-school learning centers, where youth from low-income communities learn to express themselves creatively with new technologies.
This week, Anna (https://twitter.com/annarrose) and Tarun (https://twitter.com/tarunchitra) chat with Florian Tramèr (https://twitter.com/florian_tramer), Assistant Professor at ETH Zurich (https://ethz.ch/en.html). They discuss his earlier work on side channel attacks on privacy blockchains, as well as his academic focus on Machine Learning (ML) and adversarial research. They define some key ML terms, tease out some of the nuances of ML training and models, chat zkML and other privacy environments where ML can be trained, and look at why the security around ML will be important as these models become increasingly used in production. Here are some additional links for this episode: * Episode 228: Catch-up at DevConnect AMS with Tarun, Guillermo and Brendan (https://zeroknowledge.fm/228a/) * Florian Tramèr's Github (https://github.com/ftramer) * Florian Tramèr's Publications & Papers (https://floriantramer.com/publications/) * ETH Zurich (https://ethz.ch/en.html) * DevConnect (https://devconnect.org/) * Tarun Chritra's Github (https://github.com/pluriholonomic) * Single Secret Leader Election by Dan Boneh, Saba Eskandarian, Lucjan Hanzlik, and Nicola Greco (https://eprint.iacr.org/2020/025) * GasToken: A Journey Through Blockchain Resource Arbitrage by Tramèr, Daian, Breidenbach and Juels (https://floriantramer.com/docs/slides/CESC18gastoken.pdf) * Enter the Hydra: Towards Principled Bug Bounties and Exploit-Resistant Smart Contracts by Tramèr, Daian, Breidenbach and Juels (https://eprint.iacr.org/2017/1090) * Ronin Bridge Hack – Community Alert: Ronin Validators Compromised (https://roninblockchain.substack.com/p/community-alert-ronin-validators?s=w) * InstaHide: Instance-hiding Schemes for Private Distributed Learning, Huang et al. 2020. (https://arxiv.org/abs/2010.02772) * Is Private Learning Possible with Instance Encoding? (https://arxiv.org/abs/2011.05315) * OpenAI's GPT-3 model (https://openai.com/api/) * OpenAI's GPT-2 model (https://openai.com/blog/tags/gpt-2/) * OpenAI's GPT-2 model (https://openai.com/blog/tags/gpt-2/) * The Part-Time Parliament, Lamport, 1998. (https://lamport.azurewebsites.net/pubs/lamport-paxos.pdf) * You Autocomplete Me: Poisoning Vulnerabilities in Neural Code Completion (https://arxiv.org/abs/2007.02220) ZK Whiteboard Sessions (https://zkhack.dev/whiteboard/) – as part of ZK Hack and powered by Polygon – a new series of educational videos that will help you get onboarded into the concepts and terms that we talk about on the ZK front. ZK Jobs Board (https://jobsboard.zeroknowledge.fm/) – has a fresh batch of open roles from ZK-focused projects. Find your next opportunity working in ZK! Today's episode is sponsored by Mina Protocol (https://minaprotocol.com/). With Mina's zero knowledge smart contracts – or zkApps – developers can create apps that offer privacy, security, and verifiability for your users. Head to minaprotocol.com/zkpodcast (http://minaprotocol.com/zkpodcast) to learn about their developer bootcamps and open grants. If you like what we do: * Find all our links here! @ZeroKnowledge | Linktree (https://linktr.ee/zeroknowledge) * Subscribe to our podcast newsletter (https://zeroknowledge.substack.com) * Follow us on Twitter @zeroknowledgefm (https://twitter.com/zeroknowledgefm) * Join us on Telegram (https://zeroknowledge.fm/telegram) * Catch us on Youtube (https://zeroknowledge.fm/) * Head to the ZK Community Forum (https://community.zeroknowledge.fm/) * Support our Gitcoin Grant (https://zeroknowledge.fm/gitcoin-grant-329-zkp-2)
Send us a textThis is the second part of my deep dive into the Lucas Research Study on Project Based Learning. You can find the first part in episode 88. The research findings absolutely dispel the myth that PBL doesn't improve test scores. Project Based Learning improves AP test scores for high school students and creates noticeable gains in middle school and elementary school students.It also increased engagement for the teachers. Having a culture with sustained professional learning and coaching for teachers and staff creates an environment where PBL is done right and increases outcomes and enjoyment for learners and teachers. The study shows that PBL was beneficial for all learners and staff. Episode Highlights: [02:10] "Need to Know" - How does our PBL school get better throughout the school year? [02:50] Having a staff member or PBL coach that specifically looks at Project Based Learning and supports the staff is essential. [04:51] Administrators also need an administrative coach. You need someone outside of your staff and district to push you.[06:04] Project Based Learning boosts student learning in AP courses.[07:48] Many of the participants were from underserved communities, and the study was created to help promote equity. The program increased test performance when using Project Based Learning.[08:31] The myth that Project Based Learning does not enhance test scores has been flipped on its head. [09:37] Your staff has to have professional learning in order to implement PBL properly.[10:24] PBL also led to gains in math, science, and vocabulary learning for middle schoolers. The teachers were also more engaged.[14:48] PBL is beneficial for all students.Resources & Links Related to this EpisodeWhat is PBL?Ask RyanMagnify Learning YouTubeProject Based Learning Stories and Structures: Wins, Fails, and Where to StartMagnify LearningRyan Steuer Twitter @ryansteuerPBL Workshops For Schools & DistrictsCommunity Partner ResourcesLucas Research Briefs on Project Based LearningPBL Leadership - Project Based Learning Research Part 1 | Episode 88PBL Guest - Danny Bauer of Better Leaders Better Schools | Episode 87
This is the second part of my deep dive into the Lucas Research Study on Project Based Learning. You can find the first part in episode 88. The research findings absolutely dispel the myth that PBL doesn't improve test scores. Project Based Learning improves AP test scores for high school students and creates noticeable gains in middle school and elementary school students.It also increased engagement for the teachers. Having a culture with sustained professional learning and coaching for teachers and staff creates an environment where PBL is done right and increases outcomes and enjoyment for learners and teachers. The study shows that PBL was beneficial for all learners and staff. Episode Highlights: [02:10] "Need to Know" - How does our PBL school get better throughout the school year? [02:50] Having a staff member or PBL coach that specifically looks at Project Based Learning and supports the staff is essential. [04:51] Administrators also need an administrative coach. You need someone outside of your staff and district to push you.[06:04] Project Based Learning boosts student learning in AP courses.[07:48] Many of the participants were from underserved communities, and the study was created to help promote equity. The program increased test performance when using Project Based Learning.[08:31] The myth that Project Based Learning does not enhance test scores has been flipped on its head. [09:37] Your staff has to have professional learning in order to implement PBL properly.[10:24] PBL also led to gains in math, science, and vocabulary learning for middle schoolers. The teachers were also more engaged.[14:48] PBL is beneficial for all students.Resources & Links Related to this EpisodeWhat is PBL?Ask RyanMagnify Learning YouTubeProject Based Learning Stories and Structures: Wins, Fails, and Where to StartMagnify LearningRyan Steuer Twitter @ryansteuerPBL Workshops For Schools & DistrictsCommunity Partner ResourcesLucas Research Briefs on Project Based LearningPBL Leadership - Project Based Learning Research Part 1 | Episode 88PBL Guest - Danny Bauer of Better Leaders Better Schools | Episode 87
Send us a textI've seen lives turned around by Project Based Learning. I have loads of success stories from learners, teachers, and administrators. There is also research to back up the benefits of Project Based Learning. This leadership episode is going to be a two-part episode focusing on Research on Project Based Learning by Lucas Education Research. The brief is linked to in the resources below.This episode takes a high-level view, and then I break it down in the upcoming episode 90 as we take a deeper look. We also have a “need to know” that focuses on the question of where do I get started with PBL? I share a free resource in the show notes and other options for learning about Project Based Learning. I also share the best way to learn by visiting a school and seeing it in action. Episode Highlights: [02:54] Where do I get started with PBL? Educate yourself with books, podcasts, videos, and visiting a school that is doing PBL.[04:01] On your visit ask good questions and be sure to talk to the learners.[05:55] It's a powerful lever for improving equity. When it's used significant learning occurs. [07:05] The study shows that when underserved students engage in PBL they learn significantly.[09:22] Learners miss out on opportunities and authentic learning experiences when not involved in Project Based Learning.[11:59] Curriculum needs to be flexible enough to pull in additional resources to help your students connect it to the work. Connections are a huge benefit.[13:06] Belonging in the school community. We need to build the culture so students are in a safe environment.[15:37] PBL also creates strong learning opportunities for teachers. Sustained high-quality professional learning. Resources & Links Related to this EpisodeWhat is PBL?Ask RyanMagnify Learning YouTubeProject Based Learning Stories and Structures: Wins, Fails, and Where to StartMagnify LearningRyan Steuer Twitter @ryansteuerPBL Workshops For Schools & DistrictsCommunity Partner ResourcesLucas Research Briefs on Project Based Learning
I've seen lives turned around by Project Based Learning. I have loads of success stories from learners, teachers, and administrators. There is also research to back up the benefits of Project Based Learning. This leadership episode is going to be a two-part episode focusing on Research on Project Based Learning by Lucas Education Research. The brief is linked to in the resources below.This episode takes a high-level view, and then I break it down in the upcoming episode 90 as we take a deeper look. We also have a “need to know” that focuses on the question of where do I get started with PBL? I share a free resource in the show notes and other options for learning about Project Based Learning. I also share the best way to learn by visiting a school and seeing it in action. Episode Highlights: [02:54] Where do I get started with PBL? Educate yourself with books, podcasts, videos, and visiting a school that is doing PBL.[04:01] On your visit ask good questions and be sure to talk to the learners.[05:55] It's a powerful lever for improving equity. When it's used significant learning occurs. [07:05] The study shows that when underserved students engage in PBL they learn significantly.[09:22] Learners miss out on opportunities and authentic learning experiences when not involved in Project Based Learning.[11:59] Curriculum needs to be flexible enough to pull in additional resources to help your students connect it to the work. Connections are a huge benefit.[13:06] Belonging in the school community. We need to build the culture so students are in a safe environment.[15:37] PBL also creates strong learning opportunities for teachers. Sustained high-quality professional learning. Resources & Links Related to this EpisodeWhat is PBL?Ask RyanMagnify Learning YouTubeProject Based Learning Stories and Structures: Wins, Fails, and Where to StartMagnify LearningRyan Steuer Twitter @ryansteuerPBL Workshops For Schools & DistrictsCommunity Partner ResourcesLucas Research Briefs on Project Based Learning
If you have a child (a young sibling, cousin, student, or even friend) in your life, chances are you know about Scratch—the wildly popular graphical program language that kids use to dream up interactive stories, games, and animations. But it's entirely possible you don't know the man behind Scratch, Mitch Resnick, the LEGO Papert Professor of Learning Research at the MIT Media Lab, or that the Scratch Foundation has a strong, long-term relationship with EPAM. After listening to the latest iteration of *The Resonance Test,* in which Resnick and Shamilka Samarasinha, EPAM's Global Head of Corporate Social Responsibility, answer questions from producer Ken Gordon, that ignorance will instantly evaporate. The episode digs into the reasons why Scratch is, and always has been, a free program (“We didn't want there to be barriers for young people to get access to Scratch,” says Resnick) and how Scratch helped children during Covid (the first year of the pandemic saw the number of Scratch projects double and the number of comments the kids wrote on each other's projects increase fivefold). You'll hear about Scratch and the kids of Ukraine. Says Resnick: “In early March, 10 days after the invasion of Ukraine, I got a message from an educator in Ukraine named Olesia Vlasii.” Vlasii had the idea to use Scratch to create what she called Waves of Kindness. The result: A Waves of Kindness gallery on the Scratch website “where kids from around the world could upload projects about how you could spread kindness,” says Resnick. Within days, Waves of Kindness featured “literally thousands of projects from kids around the world.” The conversation also touches on how Scratch engages a wide ecosystem of learners to promote diversity and inclusion in expanding education and how EPAM's partnership with Scratch fits into our other ESG activities. “In the social impact space, obviously education is one of our core areas,” says Samarasinha. Finally, Resnik and Samarasinha talk about the evolving relationship between the Scratch Foundation and EPAM—our EPAM E-Kids program has expanded from four to 19 countries—and the upcoming virtual Scratch Conference. It's a conversation that your kid will want you to hear. So listen! Host: Alison Kotin Engineer: Kyp Pilalas Producer: Ken Gordon
Vin Vashishta is a chief data officer and AI strategist at V Squared, a company he founded in 2012 that provides AI strategy, transformation, and data organizational build-out services. He teaches data professionals about strategy, communications, business acumen, and applied machine learning research methods. Vin has 130k+ followers on Linkedin talking about AI, analytics, and strategy. His website: https://www.datascience.vin/ Follow @DalianaLiu for more updates on data science and this show. If you enjoy this episode, subscribe and leave a 5-star review :) Topics: Machine learning problem solving Case study: how to find pricing strategy through data science Applied ML research methods “Human machine team” Casual inference resources/books How to earn trust from customer Future of data science/Machine learning MLOps vs QA (quality control) How to lead without authority Mistakes he made in his career What he learned from his mentor Shift in data science over the passed 10 years
Adam and Kathryn talk about Pilates, yoga and how we can integrate current evidence into the practices we love without throwing the baby out with the bath water. Adam is on his way to becoming a physical therapist and along the path has incorporated heavy weights into his long-time Pilates career. They also talk about motor learning and how we can cue our clients to be as successful as possible during exercise. ___ If this conversation excites you click here to learn more about the Mindful Strength Teacher's Immersion 2.0. Early bird ends July 1st. This 3-month continuing education course for yoga, pilates and fitness teachers begins in October. ___ Adam lives in Long Beach, California and has been teaching Pilates since 2009. He is an internationally recognized Pilates teacher trainer for Breathe Education, holds a Bachelors of Science in Exercise Science and is currently studying Physical Therapy at the University of St. Augustine for Health Sciences. Adam is the founder of various live online workshops such as “Motor Learning Strategies” & “Pain Science & Pilates To learn more about Adam click here. To follow Adam on IG click here.
#dsi #search #google Search engines work by building an index and then looking up things in it. Usually, that index is a separate data structure. In keyword search, we build and store reverse indices. In neural search, we build nearest-neighbor indices. This paper does something different: It directly trains a Transformer to return the ID of the most relevant document. No similarity search over embeddings or anything like this is performed, and no external data structure is needed, as the entire index is essentially captured by the model's weights. The paper experiments with various ways of representing documents and training the system, which works surprisingly well! Sponsor: Diffgram https://diffgram.com?ref=yannic OUTLINE: 0:00 - Intro 0:45 - Sponsor: Diffgram 1:35 - Paper overview 3:15 - The search problem, classic and neural 8:15 - Seq2seq for directly predicting document IDs 11:05 - Differentiable search index architecture 18:05 - Indexing 25:15 - Retrieval and document representation 33:25 - Training DSI 39:15 - Experimental results 49:25 - Comments & Conclusions Paper: https://arxiv.org/abs/2202.06991 Abstract: In this paper, we demonstrate that information retrieval can be accomplished with a single Transformer, in which all information about the corpus is encoded in the parameters of the model. To this end, we introduce the Differentiable Search Index (DSI), a new paradigm that learns a text-to-text model that maps string queries directly to relevant docids; in other words, a DSI model answers queries directly using only its parameters, dramatically simplifying the whole retrieval process. We study variations in how documents and their identifiers are represented, variations in training procedures, and the interplay between models and corpus sizes. Experiments demonstrate that given appropriate design choices, DSI significantly outperforms strong baselines such as dual encoder models. Moreover, DSI demonstrates strong generalization capabilities, outperforming a BM25 baseline in a zero-shot setup. Authors: Yi Tay, Vinh Q. Tran, Mostafa Dehghani, Jianmo Ni, Dara Bahri, Harsh Mehta, Zhen Qin, Kai Hui, Zhe Zhao, Jai Gupta, Tal Schuster, William W. Cohen, Donald Metzler Links: Merch: store.ykilcher.com TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... LinkedIn: https://www.linkedin.com/in/ykilcher BiliBili: https://space.bilibili.com/2017636191 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
Today I had the pleasure of interviewing Yannic Kilcher. Yannic is a YouTuber covering state of the art Machine Learning research topics. He has a PhD from ETH Zurich and is currently the CTO of DeepJudge, a LegalTech NLP startup. In this episode we learn about how Yannic decided on a PHD in Ai, how he is able to make advanced research so digestable, and the reason why he wears sunglasses on camera. I hope you enjoy the epsisode, I know I enjoyed our conversation.
“Why is the sky blue?” “Why do people get sick?” “Why aren't there any more dinosaurs?” Sometimes it feels like children never stop asking questions. And they shouldn't. A recent OECD International Early Learning and Child Wellbeing study shows that children who are curious have stronger language and number skills, and better self-control. So how do we keep students curious and creative even after they've outgrown kindergarten? Rowena Phair, senior analyst at the OECD, and Mitch Resnick, Professor of Learning Research at the MIT Media Lab, discuss. Host: Clara Young; Producer: Ilse Sánchez
Mathilde Caron is a PhD. candidate at the French National Institute for Research in Digital Science and Technology and at Facebook AI (Meta AI). She does the majority of her research in the field of Machine learning called self-supervised learning. She has a few first authorships on important academic papers in the space. Her work: https://scholar.google.com/citations?user=eiB0s-kAAAAJ&hl=fr You can donate to this podcast at this bitcoin address: 33wejXuGGDtQj9GPwCgjwPxPq4dc4muZjg --- This episode is sponsored by · Anchor: The easiest way to make a podcast. https://anchor.fm/app Support this podcast: https://anchor.fm/idris-sunmola/support
#grafting #adam #sgd The last years in deep learning research have given rise to a plethora of different optimization algorithms, such as SGD, AdaGrad, Adam, LARS, LAMB, etc. which all claim to have their special peculiarities and advantages. In general, all algorithms modify two major things: The (implicit) learning rate schedule, and a correction to the gradient direction. This paper introduces grafting, which allows to transfer the induced learning rate schedule of one optimizer to another one. In that, the paper shows that much of the benefits of adaptive methods (e.g. Adam) are actually due to this schedule, and not necessarily to the gradient direction correction. Grafting allows for more fundamental research into differences and commonalities between optimizers, and a derived version of it makes it possible to computes static learning rate corrections for SGD, which potentially allows for large savings of GPU memory. OUTLINE 0:00 - Rant about Reviewer #2 6:25 - Intro & Overview 12:25 - Adaptive Optimization Methods 20:15 - Grafting Algorithm 26:45 - Experimental Results 31:35 - Static Transfer of Learning Rate Ratios 35:25 - Conclusion & Discussion Paper (OpenReview): https://openreview.net/forum?id=FpKgG... Old Paper (Arxiv): https://arxiv.org/abs/2002.11803 Our Discord: https://discord.gg/4H8xxDF Abstract: In the empirical science of training large neural networks, the learning rate schedule is a notoriously challenging-to-tune hyperparameter, which can depend on all other properties (architecture, optimizer, batch size, dataset, regularization, ...) of the problem. In this work, we probe the entanglements between the optimizer and the learning rate schedule. We propose the technique of optimizer grafting, which allows for the transfer of the overall implicit step size schedule from a tuned optimizer to a new optimizer, preserving empirical performance. This provides a robust plug-and-play baseline for optimizer comparisons, leading to reductions to the computational cost of optimizer hyperparameter search. Using grafting, we discover a non-adaptive learning rate correction to SGD which allows it to train a BERT model to state-of-the-art performance. Besides providing a resource-saving tool for practitioners, the invariances discovered via grafting shed light on the successes and failure modes of optimizers in deep learning. Authors: Anonymous (Under Review) Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... LinkedIn: https://www.linkedin.com/in/ykilcher BiliBili: https://space.bilibili.com/2017636191 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#deeplearning #backpropagation #simulation More and more systems are made differentiable, which means that accurate gradients of these systems' dynamics can be computed exactly. While this development has led to a lot of advances, there are also distinct situations where backpropagation can be a very bad idea. This paper characterizes a few such systems in the domain of iterated dynamical systems, often including some source of stochasticity, resulting in chaotic behavior. In these systems, it is often better to use black-box estimators for gradients than computing them exactly. OUTLINE: 0:00 - Foreword 1:15 - Intro & Overview 3:40 - Backpropagation through iterated systems 12:10 - Connection to the spectrum of the Jacobian 15:35 - The Reparameterization Trick 21:30 - Problems of reparameterization 26:35 - Example 1: Policy Learning in Simulation 33:05 - Example 2: Meta-Learning Optimizers 36:15 - Example 3: Disk packing 37:45 - Analysis of Jacobians 40:20 - What can be done? 45:40 - Just use Black-Box methods Paper: https://arxiv.org/abs/2111.05803 Abstract: Differentiable programming techniques are widely used in the community and are responsible for the machine learning renaissance of the past several decades. While these methods are powerful, they have limits. In this short report, we discuss a common chaos based failure mode which appears in a variety of differentiable circumstances, ranging from recurrent neural networks and numerical physics simulation to training learned optimizers. We trace this failure to the spectrum of the Jacobian of the system under study, and provide criteria for when a practitioner might expect this failure to spoil their differentiation based optimization algorithms. Authors: Luke Metz, C. Daniel Freeman, Samuel S. Schoenholz, Tal Kachman Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... LinkedIn: https://www.linkedin.com/in/ykilcher BiliBili: https://space.bilibili.com/2017636191 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#machinelearning #ardm #generativemodels Diffusion models have made large advances in recent months as a new type of generative models. This paper introduces Autoregressive Diffusion Models (ARDMs), which are a mix between autoregressive generative models and diffusion models. ARDMs are trained to be agnostic to the order of autoregressive decoding and give the user a dynamic tradeoff between speed and performance at decoding time. This paper applies ARDMs to both text and image data, and as an extension, the models can also be used to perform lossless compression. OUTLINE: 0:00 - Intro & Overview 3:15 - Decoding Order in Autoregressive Models 6:15 - Autoregressive Diffusion Models 8:35 - Dependent and Independent Sampling 14:25 - Application to Character-Level Language Models 18:15 - How Sampling & Training Works 26:05 - Extension 1: Parallel Sampling 29:20 - Extension 2: Depth Upscaling 33:10 - Conclusion & Comments Paper: https://arxiv.org/abs/2110.02037 Abstract: We introduce Autoregressive Diffusion Models (ARDMs), a model class encompassing and generalizing order-agnostic autoregressive models (Uria et al., 2014) and absorbing discrete diffusion (Austin et al., 2021), which we show are special cases of ARDMs under mild assumptions. ARDMs are simple to implement and easy to train. Unlike standard ARMs, they do not require causal masking of model representations, and can be trained using an efficient objective similar to modern probabilistic diffusion models that scales favourably to highly-dimensional data. At test time, ARDMs support parallel generation which can be adapted to fit any given generation budget. We find that ARDMs require significantly fewer steps than discrete diffusion models to attain the same performance. Finally, we apply ARDMs to lossless compression, and show that they are uniquely suited to this task. Contrary to existing approaches based on bits-back coding, ARDMs obtain compelling results not only on complete datasets, but also on compressing single data points. Moreover, this can be done using a modest number of network calls for (de)compression due to the model's adaptable parallel generation. Authors: Emiel Hoogeboom, Alexey A. Gritsenko, Jasmijn Bastings, Ben Poole, Rianne van den Berg, Tim Salimans Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/ykilcher BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m
#efficientzero #muzero #atari Reinforcement Learning methods are notoriously data-hungry. Notably, MuZero learns a latent world model just from scalar feedback of reward- and policy-predictions, and therefore relies on scale to perform well. However, most RL algorithms fail when presented with very little data. EfficientZero makes several improvements over MuZero that allows it to learn from astonishingly small amounts of data and outperform other methods by a large margin in the low-sample setting. This could be a staple algorithm for future RL research. OUTLINE: 0:00 - Intro & Outline 2:30 - MuZero Recap 10:50 - EfficientZero improvements 14:15 - Self-Supervised consistency loss 17:50 - End-to-end prediction of the value prefix 20:40 - Model-based off-policy correction 25:45 - Experimental Results & Conclusion Paper: https://arxiv.org/abs/2111.00210 Code: https://github.com/YeWR/EfficientZero Note: code not there yet as of release of this video Abstract: Reinforcement learning has achieved great success in many applications. However, sample efficiency remains a key challenge, with prominent methods requiring millions (or even billions) of environment steps to train. Recently, there has been significant progress in sample efficient image-based RL algorithms; however, consistent human-level performance on the Atari game benchmark remains an elusive goal. We propose a sample efficient model-based visual RL algorithm built on MuZero, which we name EfficientZero. Our method achieves 190.4% mean human performance and 116.0% median performance on the Atari 100k benchmark with only two hours of real-time game experience and outperforms the state SAC in some tasks on the DMControl 100k benchmark. This is the first time an algorithm achieves super-human performance on Atari games with such little data. EfficientZero's performance is also close to DQN's performance at 200 million frames while we consume 500 times less data. EfficientZero's low sample complexity and high performance can bring RL closer to real-world applicability. We implement our algorithm in an easy-to-understand manner and it is available at this https URL. We hope it will accelerate the research of MCTS-based RL algorithms in the wider community. Authors: Weirui Ye, Shaohuai Liu, Thanard Kurutach, Pieter Abbeel, Yang Gao Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/ykilcher BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
BIO: Andre Hsu is a thinker and business strategist based in Singapore. He is the author of three books about qualities, mindsets, and frameworks relevant to business, which he observed in several business tycoons who left a deep and lasting impact on his life. STORY: Andre partnered with a software company with an excellent business idea, but the partners were poor in managing the company and selling the product, so it failed. Andre lost his entire investment. LEARNING: Research the people who own a business as much as you research the business. “No matter how great the idea is, you'll not succeed if you cannot trust your partner.”Andre Hsu Guest profilehttps://www.linkedin.com/in/andrehsujs/ (Andre Hsu) is a thinker and business strategist based in Singapore. He started his entrepreneurial journey at 17, working on a real estate project while juggling three academic degrees completed concurrently in Australia. He is the author of three books about qualities, mindsets, and frameworks that are relevant for business which he observed in several business tycoons who left a deep and lasting impact on his life. Andre likes to use multiple techniques to read, predict people, assess situations, and formulate strategies that are suitable for properties and negotiations, deal structuring related to the assets. He likes to educate, share knowledge, insights, and reasoning of strategies to business associates who then proceed to implement them. He looks forward to doing this with people who share similar values and visions. Worst investment everAfter finishing university, Andre went into the family business, and after a while, he decided to venture into his own business pursuits. He got involved in a small software company that was dealing with Point-of-Service systems. At the time, this was a very lucrative business because not many companies were using POS systems. Andre was, therefore, happy to partner with the company and invest in this venture. The mistake Andre made was investing in people who didn't take the business part of the venture seriously. They only created a good product, but they never invested in sales or management, so the product never really took off. Lessons learnedYou need to research the people who own a business as much as you research the business. Andrew's takeawaysWhen investing in a startup, you need to look for trust, a good idea, the ability to execute the idea, and capital. Actionable adviceIf you cannot trust the person you want to partner with, forget the idea. No matter how great the idea is, you'll not succeed if you cannot trust your partner. No. 1 goal for the next 12 monthsAndre Hsu's number one goal for the next 12 months is to step back from the business and have his partners run it to have more time to do strategic thinking and come up with new ideas. [spp-transcript] Connect with Andre Hsuhttps://www.linkedin.com/in/andrehsujs/ (LinkedIn) https://amzn.to/3nNgf64 (Book) Andrew's bookshttps://amzn.to/3qrfHjX (How to Start Building Your Wealth Investing in the Stock Market) https://amzn.to/2PDApAo (My Worst Investment Ever) https://amzn.to/3v6ip1Y (9 Valuation Mistakes and How to Avoid Them) https://amzn.to/3emBO8M (Transform Your Business with Dr.Deming's 14 Points) Andrew's online programshttps://valuationmasterclass.com/ (Valuation Master Class) https://academy.astotz.com/courses/how-to-start-building-your-wealth-investing-in-the-stock-market (How to Start Building Your Wealth Investing in the Stock Market) https://academy.astotz.com/courses/finance-made-ridiculously-simple (Finance Made Ridiculously Simple) https://academy.astotz.com/courses/gp (Become a Great Presenter and Increase Your Influence) https://academy.astotz.com/courses/transformyourbusiness (Transform Your Business with Dr. Deming's 14 Points) Connect with Andrew Stotz:https://www.astotz.com/ (astotz.com) https://www.linkedin.com/in/andrewstotz/ (LinkedIn)...
#tvae #topographic #equivariant Variational Autoencoders model the latent space as a set of independent Gaussian random variables, which the decoder maps to a data distribution. However, this independence is not always desired, for example when dealing with video sequences, we know that successive frames are heavily correlated. Thus, any latent space dealing with such data should reflect this in its structure. Topographic VAEs are a framework for defining correlation structures among the latent variables and induce equivariance within the resulting model. This paper shows how such correlation structures can be built by correctly arranging higher-level variables, which are themselves independent Gaussians. OUTLINE: 0:00 - Intro 1:40 - Architecture Overview 6:30 - Comparison to regular VAEs 8:35 - Generative Mechanism Formulation 11:45 - Non-Gaussian Latent Space 17:30 - Topographic Product of Student-t 21:15 - Introducing Temporal Coherence 24:50 - Topographic VAE 27:50 - Experimental Results 31:15 - Conclusion & Comments Paper: https://arxiv.org/abs/2109.01394 Code: https://github.com/akandykeller/topog... Abstract: In this work we seek to bridge the concepts of topographic organization and equivariance in neural networks. To accomplish this, we introduce the Topographic VAE: a novel method for efficiently training deep generative models with topographically organized latent variables. We show that such a model indeed learns to organize its activations according to salient characteristics such as digit class, width, and style on MNIST. Furthermore, through topographic organization over time (i.e. temporal coherence), we demonstrate how predefined latent space transformation operators can be encouraged for observed transformed input sequences -- a primitive form of unsupervised learned equivariance. We demonstrate that this model successfully learns sets of approximately equivariant features (i.e. "capsules") directly from sequences and achieves higher likelihood on correspondingly transforming test sequences. Equivariance is verified quantitatively by measuring the approximate commutativity of the inference network and the sequence transformations. Finally, we demonstrate approximate equivariance to complex transformations, expanding upon the capabilities of existing group equivariant neural networks. Authors: T. Anderson Keller, Max Welling Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/ykilcher BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#attention #transformer #fastformer Transformers have become the dominant model class in the last few years for large data, but their quadratic complexity in terms of sequence length has plagued them until now. Fastformer claims to be the fastest and most performant linear attention variant, able to consume long contexts at once. This is achieved by a combination of additive attention and elementwise products. While initial results look promising, I have my reservations... OUTLINE: 0:00 - Intro & Outline 2:15 - Fastformer description 5:20 - Baseline: Classic Attention 10:00 - Fastformer architecture 12:50 - Additive Attention 18:05 - Query-Key element-wise multiplication 21:35 - Redundant modules in Fastformer 25:00 - Problems with the architecture 27:30 - Is this even attention? 32:20 - Experimental Results 34:50 - Conclusion & Comments Paper: https://arxiv.org/abs/2108.09084 Abstract: Transformer is a powerful model for text understanding. However, it is inefficient due to its quadratic complexity to input sequence length. Although there are many methods on Transformer acceleration, they are still either inefficient on long sequences or not effective enough. In this paper, we propose Fastformer, which is an efficient Transformer model based on additive attention. In Fastformer, instead of modeling the pair-wise interactions between tokens, we first use additive attention mechanism to model global contexts, and then further transform each token representation based on its interaction with global context representations. In this way, Fastformer can achieve effective context modeling with linear complexity. Extensive experiments on five datasets show that Fastformer is much more efficient than many existing Transformer models and can meanwhile achieve comparable or even better long text modeling performance. Authors: Chuhan Wu, Fangzhao Wu, Tao Qi, Yongfeng Huang Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#pondernet #deepmind #machinelearning Humans don't spend the same amount of mental effort on all problems equally. Instead, we respond quickly to easy tasks, and we take our time to deliberate hard tasks. DeepMind's PonderNet attempts to achieve the same by dynamically deciding how many computation steps to allocate to any single input sample. This is done via a recurrent architecture and a trainable function that computes a halting probability. The resulting model performs well in dynamic computation tasks and is surprisingly robust to different hyperparameter settings. OUTLINE: 0:00 - Intro & Overview 2:30 - Problem Statement 8:00 - Probabilistic formulation of dynamic halting 14:40 - Training via unrolling 22:30 - Loss function and regularization of the halting distribution 27:35 - Experimental Results 37:10 - Sensitivity to hyperparameter choice 41:15 - Discussion, Conclusion, Broader Impact Paper: https://arxiv.org/abs/2107.05407 Abstract: In standard neural networks the amount of computation used grows with the size of the inputs, but not with the complexity of the problem being learnt. To overcome this limitation we introduce PonderNet, a new algorithm that learns to adapt the amount of computation based on the complexity of the problem at hand. PonderNet learns end-to-end the number of computational steps to achieve an effective compromise between training prediction accuracy, computational cost and generalization. On a complex synthetic problem, PonderNet dramatically improves performance over previous adaptive computation methods and additionally succeeds at extrapolation tests where traditional neural networks fail. Also, our method matched the current state of the art results on a real world question and answering dataset, but using less compute. Finally, PonderNet reached state of the art results on a complex task designed to test the reasoning capabilities of neural networks.1 Authors: Andrea Banino, Jan Balaguer, Charles Blundell Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
Why and where do companies fail at productionizing ML models? Watch the full podcast with Aarti here: https://youtu.be/VWJXiszQpTUAarti is a machine learning engineer at Snorkel AI. Prior to that, she worked closely with Andrew Ng in various capacities. She graduated with a master's in CS from Stanford, and bachelor's in CS and Computer Engineering from @New York University, and at @Microsoft Research as a research intern for John Langford, where she contributed to Vowpal Wabbit, an open-source project. About the Host:Jay is a Ph.D. student at Arizona State University, doing research on building Interpretable AI models for Medical Diagnosis.Jay Shah: https://www.linkedin.com/in/shahjay22/You can reach out to https://www.public.asu.edu/~jgshah1/ for any queries.Stay tuned for upcoming webinars!***Disclaimer: The information contained in this video represents the views and opinions of the speaker and does not necessarily represent the views or opinions of any institution. It does not constitute an endorsement by any Institution or its affiliates of such video content.***
#adversarialexamples #dimpledmanifold #security Adversarial Examples have long been a fascinating topic for many Machine Learning researchers. How can a tiny perturbation cause the neural network to change its output by so much? While many explanations have been proposed over the years, they all appear to fall short. This paper attempts to comprehensively explain the existence of adversarial examples by proposing a view of the classification landscape, which they call the Dimpled Manifold Model, which says that any classifier will adjust its decision boundary to align with the low-dimensional data manifold, and only slightly bend around the data. This potentially explains many phenomena around adversarial examples. Warning: In this video, I disagree. Remember that I'm not an authority, but simply give my own opinions. OUTLINE: 0:00 - Intro & Overview 7:30 - The old mental image of Adversarial Examples 11:25 - The new Dimpled Manifold Hypothesis 22:55 - The Stretchy Feature Model 29:05 - Why do DNNs create Dimpled Manifolds? 38:30 - What can be explained with the new model? 1:00:40 - Experimental evidence for the Dimpled Manifold Model 1:10:25 - Is Goodfellow's claim debunked? 1:13:00 - Conclusion & Comments Paper: https://arxiv.org/abs/2106.10151 My replication code: https://gist.github.com/yk/de8d987c4e... Goodfellow's Talk: https://youtu.be/CIfsB_EYsVI?t=4280 Abstract: The extreme fragility of deep neural networks when presented with tiny perturbations in their inputs was independently discovered by several research groups in 2013, but in spite of enormous effort these adversarial examples remained a baffling phenomenon with no clear explanation. In this paper we introduce a new conceptual framework (which we call the Dimpled Manifold Model) which provides a simple explanation for why adversarial examples exist, why their perturbations have such tiny norms, why these perturbations look like random noise, and why a network which was adversarially trained with incorrectly labeled images can still correctly classify test images. In the last part of the paper we describe the results of numerous experiments which strongly support this new model, and in particular our assertion that adversarial perturbations are roughly perpendicular to the low dimensional manifold which contains all the training examples. Abstract: Adi Shamir, Odelia Melamed, Oriel BenShmuel Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/ykilcher BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
Three research scientists from Google share their journey about interest in Machine Learning research and how they got started with it.Watch full podcasts with each of these speakers:Azalia Mirhoseini: https://youtu.be/5LCfH8YiOv4Sara Hooker: https://youtu.be/MHtbZls2utsNatasha Jacques: https://youtu.be/8XpCnmvq49sAbout the Host:Jay is a Ph.D. student at Arizona State University, doing research on building Interpretable AI models for Medical Diagnosis.Jay Shah: https://www.linkedin.com/in/shahjay22/You can reach out to https://www.public.asu.edu/~jgshah1/ for any queries.Stay tuned for upcoming webinars!***Disclaimer: The information contained in this video represents the views and opinions of the speaker and does not necessarily represent the views or opinions of any institution. It does not constitute an endorsement by any Institution or its affiliates of such video content.***
#xcit #transformer #attentionmechanism After dominating Natural Language Processing, Transformers have taken over Computer Vision recently with the advent of Vision Transformers. However, the attention mechanism's quadratic complexity in the number of tokens means that Transformers do not scale well to high-resolution images. XCiT is a new Transformer architecture, containing XCA, a transposed version of attention, reducing the complexity from quadratic to linear, and at least on image data, it appears to perform on par with other models. What does this mean for the field? Is this even a transformer? What really matters in deep learning? OUTLINE: 0:00 - Intro & Overview 3:45 - Self-Attention vs Cross-Covariance Attention (XCA) 19:55 - Cross-Covariance Image Transformer (XCiT) Architecture 26:00 - Theoretical & Engineering considerations 30:40 - Experimental Results 33:20 - Comments & Conclusion Paper: https://arxiv.org/abs/2106.09681 Code: https://github.com/facebookresearch/xcit Abstract: Following their success in natural language processing, transformers have recently shown much promise for computer vision. The self-attention operation underlying transformers yields global interactions between all tokens ,i.e. words or image patches, and enables flexible modelling of image data beyond the local interactions of convolutions. This flexibility, however, comes with a quadratic complexity in time and memory, hindering application to long sequences and high-resolution images. We propose a "transposed" version of self-attention that operates across feature channels rather than tokens, where the interactions are based on the cross-covariance matrix between keys and queries. The resulting cross-covariance attention (XCA) has linear complexity in the number of tokens, and allows efficient processing of high-resolution images. Our cross-covariance image transformer (XCiT) is built upon XCA. It combines the accuracy of conventional transformers with the scalability of convolutional architectures. We validate the effectiveness and generality of XCiT by reporting excellent results on multiple vision benchmarks, including image classification and self-supervised feature learning on ImageNet-1k, object detection and instance segmentation on COCO, and semantic segmentation on ADE20k. Authors: Alaaeldin El-Nouby, Hugo Touvron, Mathilde Caron, Piotr Bojanowski, Matthijs Douze, Armand Joulin, Ivan Laptev, Natalia Neverova, Gabriel Synnaeve, Jakob Verbeek, Hervé Jegou Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#implicitfunction #jax #autodiff Many problems in Machine Learning involve loops of inner and outer optimization. Finding update steps for the outer loop is usually difficult, because of the.need to differentiate through the inner loop's procedure over multiple steps. Such loop unrolling is very limited and constrained to very few steps. Other papers have found solutions around unrolling in very specific, individual problems. This paper proposes a unified framework for implicit differentiation of inner optimization procedures without unrolling and provides implementations that integrate seamlessly into JAX. OUTLINE: 0:00 - Intro & Overview 2:05 - Automatic Differentiation of Inner Optimizations 4:30 - Example: Meta-Learning 7:45 - Unrolling Optimization 13:00 - Unified Framework Overview & Pseudocode 21:10 - Implicit Function Theorem 25:45 - More Technicalities 28:45 - Experiments ERRATA: - Dataset Distillation is done with respect to the training set, not the validation or test set. Paper: https://arxiv.org/abs/2105.15183 Code coming soon Abstract: Automatic differentiation (autodiff) has revolutionized machine learning. It allows expressing complex computations by composing elementary ones in creative ways and removes the burden of computing their derivatives by hand. More recently, differentiation of optimization problem solutions has attracted widespread attention with applications such as optimization as a layer, and in bi-level problems such as hyper-parameter optimization and meta-learning. However, the formulas for these derivatives often involve case-by-case tedious mathematical derivations. In this paper, we propose a unified, efficient and modular approach for implicit differentiation of optimization problems. In our approach, the user defines (in Python in the case of our implementation) a function F capturing the optimality conditions of the problem to be differentiated. Once this is done, we leverage autodiff of F and implicit differentiation to automatically differentiate the optimization problem. Our approach thus combines the benefits of implicit differentiation and autodiff. It is efficient as it can be added on top of any state-of-the-art solver and modular as the optimality condition specification is decoupled from the implicit differentiation mechanism. We show that seemingly simple principles allow to recover many recently proposed implicit differentiation methods and create new ones easily. We demonstrate the ease of formulating and solving bi-level optimization problems using our framework. We also showcase an application to the sensitivity analysis of molecular dynamics. Authors: Mathieu Blondel, Quentin Berthet, Marco Cuturi, Roy Frostig, Stephan Hoyer, Felipe Llinares-López, Fabian Pedregosa, Jean-Philippe Vert Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq
#reinforcementlearning #deepmind #agi What's the most promising path to creating Artificial General Intelligence (AGI)? This paper makes the bold claim that a learning agent maximizing its reward in a sufficiently complex environment will necessarily develop intelligence as a by-product, and that Reward Maximization is the best way to move the creation of AGI forward. The paper is a mix of philosophy, engineering, and futurism, and raises many points of discussion. OUTLINE: 0:00 - Intro & Outline 4:10 - Reward Maximization 10:10 - The Reward-is-Enough Hypothesis 13:15 - Abilities associated with intelligence 16:40 - My Criticism 26:15 - Reward Maximization through Reinforcement Learning 31:30 - Discussion, Conclusion & My Comments Paper: https://www.sciencedirect.com/science... Abstract: In this article we hypothesise that intelligence, and its associated abilities, can be understood as subserving the maximisation of reward. Accordingly, reward is enough to drive behaviour that exhibits abilities studied in natural and artificial intelligence, including knowledge, learning, perception, social intelligence, language, generalisation and imitation. This is in contrast to the view that specialised problem formulations are needed for each ability, based on other signals or objectives. Furthermore, we suggest that agents that learn through trial and error experience to maximise reward could learn behaviour that exhibits most if not all of these abilities, and therefore that powerful reinforcement learning agents could constitute a solution to artificial general intelligence. Authors: David Silver, Satinder Singh, Doina Precup, Richard S. Sutton Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#fnet #attention #fourier Do we even need Attention? FNets completely drop the Attention mechanism in favor of a simple Fourier transform. They perform almost as well as Transformers, while drastically reducing parameter count, as well as compute and memory requirements. This highlights that a good token mixing heuristic could be as valuable as a learned attention matrix. OUTLINE: 0:00 - Intro & Overview 0:45 - Giving up on Attention 5:00 - FNet Architecture 9:00 - Going deeper into the Fourier Transform 11:20 - The Importance of Mixing 22:20 - Experimental Results 33:00 - Conclusions & Comments Paper: https://arxiv.org/abs/2105.03824 ADDENDUM: Of course, I completely forgot to discuss the connection between Fourier transforms and Convolutions, and that this might be interpreted as convolutions with very large kernels. Abstract: We show that Transformer encoder architectures can be massively sped up, with limited accuracy costs, by replacing the self-attention sublayers with simple linear transformations that "mix" input tokens. These linear transformations, along with simple nonlinearities in feed-forward layers, are sufficient to model semantic relationships in several text classification tasks. Perhaps most surprisingly, we find that replacing the self-attention sublayer in a Transformer encoder with a standard, unparameterized Fourier Transform achieves 92% of the accuracy of BERT on the GLUE benchmark, but pre-trains and runs up to seven times faster on GPUs and twice as fast on TPUs. The resulting model, which we name FNet, scales very efficiently to long inputs, matching the accuracy of the most accurate "efficient" Transformers on the Long Range Arena benchmark, but training and running faster across all sequence lengths on GPUs and relatively shorter sequence lengths on TPUs. Finally, FNet has a light memory footprint and is particularly efficient at smaller model sizes: for a fixed speed and accuracy budget, small FNet models outperform Transformer counterparts. Authors: James Lee-Thorp, Joshua Ainslie, Ilya Eckstein, Santiago Ontanon Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
During this past year, one of the biggest challenges educators reported was keeping students engaged and motivated while learning online during the pandemic. In this episode of BRIGHT, we talk to Dr. Chris Harrington, the director of the Michigan Virtual Learning Research Institute, who shares his reflections on pandemic teaching, common misconceptions about online learning, and the findings of his team's landmark study on student engagement.
#ddpm #diffusionmodels #openai GANs have dominated the image generation space for the majority of the last decade. This paper shows for the first time, how a non-GAN model, a DDPM, can be improved to overtake GANs at standard evaluation metrics for image generation. The produced samples look amazing and other than GANs, the new model has a formal probabilistic foundation. Is there a future for GANs or are Diffusion Models going to overtake them for good? OUTLINE: 0:00 - Intro & Overview 4:10 - Denoising Diffusion Probabilistic Models 11:30 - Formal derivation of the training loss 23:00 - Training in practice 27:55 - Learning the covariance 31:25 - Improving the noise schedule 33:35 - Reducing the loss gradient noise 40:35 - Classifier guidance 52:50 - Experimental Results Paper (this): https://arxiv.org/abs/2105.05233 Paper (previous): https://arxiv.org/abs/2102.09672 Code: https://github.com/openai/guided-diff... Abstract: We show that diffusion models can achieve image sample quality superior to the current state-of-the-art generative models. We achieve this on unconditional image synthesis by finding a better architecture through a series of ablations. For conditional image synthesis, we further improve sample quality with classifier guidance: a simple, compute-efficient method for trading off diversity for sample quality using gradients from a classifier. We achieve an FID of 2.97 on ImageNet 128×128, 4.59 on ImageNet 256×256, and 7.72 on ImageNet 512×512, and we match BigGAN-deep even with as few as 25 forward passes per sample, all while maintaining better coverage of the distribution. Finally, we find that classifier guidance combines well with upsampling diffusion models, further improving FID to 3.85 on ImageNet 512×512. We release our code at this https URL Authors: Alex Nichol, Prafulla Dhariwal Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#mixer #google #imagenet Convolutional Neural Networks have dominated computer vision for nearly 10 years, and that might finally come to an end. First, Vision Transformers (ViT) have shown remarkable performance, and now even simple MLP-based models reach competitive accuracy, as long as sufficient data is used for pre-training. This paper presents MLP-Mixer, using MLPs in a particular weight-sharing arrangement to achieve a competitive, high-throughput model and it raises some interesting questions about the nature of learning and inductive biases and their interaction with scale for future research. OUTLINE: 0:00 - Intro & Overview 2:20 - MLP-Mixer Architecture 13:20 - Experimental Results 17:30 - Effects of Scale 24:30 - Learned Weights Visualization 27:25 - Comments & Conclusion Paper: https://arxiv.org/abs/2105.01601 Abstract: Convolutional Neural Networks (CNNs) are the go-to model for computer vision. Recently, attention-based networks, such as the Vision Transformer, have also become popular. In this paper we show that while convolutions and attention are both sufficient for good performance, neither of them are necessary. We present MLP-Mixer, an architecture based exclusively on multi-layer perceptrons (MLPs). MLP-Mixer contains two types of layers: one with MLPs applied independently to image patches (i.e. "mixing" the per-location features), and one with MLPs applied across patches (i.e. "mixing" spatial information). When trained on large datasets, or with modern regularization schemes, MLP-Mixer attains competitive scores on image classification benchmarks, with pre-training and inference cost comparable to state-of-the-art models. We hope that these results spark further research beyond the realms of well established CNNs and Transformers. Authors: Ilya Tolstikhin, Neil Houlsby, Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Thomas Unterthiner, Jessica Yung, Daniel Keysers, Jakob Uszkoreit, Mario Lucic, Alexey Dosovitskiy ERRATA: Here is their definition of what the 5-shot classifier is: "we report the few-shot accuracies obtained by solving the L2-regularized linear regression problem between the frozen learned representations of images and the labels" Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#universalcomputation #pretrainedtransformers #finetuning Large-scale pre-training and subsequent fine-tuning is a common recipe for success with transformer models in machine learning. However, most such transfer learning is done when a model is pre-trained on the same or a very similar modality to the final task to be solved. This paper demonstrates that transformers can be fine-tuned to completely different modalities, such as from language to vision. Moreover, they demonstrate that this can be done by freezing all attention layers, tuning less than .1% of all parameters. The paper further claims that language modeling is a superior pre-training task for such cross-domain transfer. The paper goes through various ablation studies to make its point. OUTLINE: 0:00 - Intro & Overview 2:00 - Frozen Pretrained Transformers 4:50 - Evaluated Tasks 10:05 - The Importance of Training LayerNorm 17:10 - Modality Transfer 25:10 - Network Architecture Ablation 26:10 - Evaluation of the Attention Mask 27:20 - Are FPTs Overfitting or Underfitting? 28:20 - Model Size Ablation 28:50 - Is Initialization All You Need? 31:40 - Full Model Training Overfits 32:15 - Again the Importance of Training LayerNorm 33:10 - Conclusions & Comments Paper: https://arxiv.org/abs/2103.05247 Code: https://github.com/kzl/universal-comp... Abstract: We investigate the capability of a transformer pretrained on natural language to generalize to other modalities with minimal finetuning -- in particular, without finetuning of the self-attention and feedforward layers of the residual blocks. We consider such a model, which we call a Frozen Pretrained Transformer (FPT), and study finetuning it on a variety of sequence classification tasks spanning numerical computation, vision, and protein fold prediction. In contrast to prior works which investigate finetuning on the same modality as the pretraining dataset, we show that pretraining on natural language improves performance and compute efficiency on non-language downstream tasks. In particular, we find that such pretraining enables FPT to generalize in zero-shot to these modalities, matching the performance of a transformer fully trained on these tasks. Authors: Kevin Lu, Aditya Grover, Pieter Abbeel, Igor Mordatch Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yann... Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-ki... BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannick... Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
#aiwinter #agi #embodiedcognition The AI community has gone through regular cycles of AI Springs, where rapid progress gave rise to massive overconfidence, high funding, and overpromise, followed by these promises being unfulfilled, subsequently diving into periods of disenfranchisement and underfunding, called AI Winters. This paper examines the reasons for the repeated periods of overconfidence and identifies four fallacies that people make when they see rapid progress in AI. OUTLINE: 0:00 - Intro & Overview 2:10 - AI Springs & AI Winters 5:40 - Is the current AI boom overhyped? 15:35 - Fallacy 1: Narrow Intelligence vs General Intelligence 19:40 - Fallacy 2: Hard for humans doesn't mean hard for computers 21:45 - Fallacy 3: How we call things matters 28:15 - Fallacy 4: Embodied Cognition 35:30 - Conclusion & Comments Paper: https://arxiv.org/abs/2104.12871 My Video on Shortcut Learning: https://youtu.be/D-eg7k8YSfs Abstract: Since its beginning in the 1950s, the field of artificial intelligence has cycled several times between periods of optimistic predictions and massive investment ("AI spring") and periods of disappointment, loss of confidence, and reduced funding ("AI winter"). Even with today's seemingly fast pace of AI breakthroughs, the development of long-promised technologies such as self-driving cars, housekeeping robots, and conversational companions has turned out to be much harder than many people expected. One reason for these repeating cycles is our limited understanding of the nature and complexity of intelligence itself. In this paper I describe four fallacies in common assumptions made by AI researchers, which can lead to overconfident predictions about the field. I conclude by discussing the open questions spurred by these fallacies, including the age-old challenge of imbuing machines with humanlike common sense. Authors: Melanie Mitchell Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yannic-kilcher Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-kilcher-488534136/ BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannickilcher Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
Today, Annalisa speaks with Dr. Omid Fotuhi about belonging. Omid speaks about belonging as a fundamental need. He talks about belonging in the workplace and belonging for young adults in college. Omid is a Research Psychologist who has dedicated his life to exploring and researching the processes and scientific mechanisms underlying human motivation and performance. He earned his PhD in Psychology from the University of Waterloo and later helped to co-found the College Transition Collaborative and Stanford Interventions Lab at Stanford University. He is currently the Director of Learning and Innovation at WGU Labs and a Research Associate at the Learning Research and Development Center at the University of Pittsburgh. Hosted by: Annalisa Holcombe; Edited by: Breanna Steggell.
In this episode of the Data Exchange, our special correspondent and managing editor Jenn Webb organized a mini-panel composed of myself and Simon Rodriguez, Data Research Assistant at the Center for Security and Emerging Technology (CSET) at Georgetown University. Through a series of reports and data briefs, CSET provides policymakers with data rich material to inform and guide public policy.Subscribe: Apple • Android • Spotify • Stitcher • Google • RSS.Detailed show notes can be found on The Data Exchange web site.Subscribe to The Gradient Flow Newsletter.
We're thrilled to announce our new partnership with Carleton University in Ottawa, Canada, to produce an 8-part mini series on the topic of Corrective Feedback. The series explores the area of corrective feedback through interviews with 8 scholars in the field. All interviews are conducted by students in Dr. Eva Kartchava's MA class at Carleton University as a means of assessment to connect researchers to their audience and have her students generate a greater level of understanding and investment in the research from the course. If you are interested in having a similar series produced for your class or institute, you can contact us: info@learnyourenglish.com This is episode 2 in our series. In this episode, Dr. Hossein Nassaji joins the program to discuss corrective feedback. Dr. Hossein is an award-winning scholar and Professor of Applied Linguistics in the Department of Linguistics at the University of Victoria, Victoria, BC. He has authored over 100 articles and many books. His forthcoming handbook on corrective feedback, The Cambridge Handbook of Corrective Feedback in Second Language Learning and Teaching with Eva Kartchava, is a comprehensive volume that discusses current issues and perspectives on corrective feedback and their applications to second language teaching and learning. Specifically in this episode, Dr. Nassaji tells us about: the roles corrective feedback plays in language learning how culture impacts feedback effectiveness the debate between immediate and delayed feedback written vs oral feedback and the efficacy of written feedback the what, when, why, and if of explicit & implicit feedback how teachers can learn about and implement corrective feedback in their classes *This interview was conducted by Kelsey Ulrich-Verslycken and Lana Haj Hamid Partnership with Carleton University: Throughout the series, MA students from Dr. Kartchava's class will interview leading experts in the field of corrective feedback. We thank Dr. Kartchava for joining this episode and for spearheading this initiative. For more information on this episode, this project, and those involved: view Carleton and Dr. Kartchava's website on Corrective Feedback view the LYE blog post on this episode More from Dr. Nassaji: His website Some of his prominent books: Corrective Feedback in Second Language Teaching and Learning: Research, Theory, Applications, Implication The Interactional Feedback Dimension in Instructed Second Language Learning: Linking Theory, Research, and Practice Teaching Grammar in Second Language Classrooms: Integrating Form-Focused Instruction in Communicative Context The Cambridge Handbook of Corrective Feedback in Second Language Learning and Teaching Podcast Creation: This episode was created with support from Thinkific & Podbean. If you're looking to launch a course or start a podcast, we highly recommend them - and use them ourselves. As always, thank you for listening. Your support has been overwhelming and we couldn't do what we do without you. We hope this podcast serves as an effective CPD tool for you. If you have a comment or question about today's show, we'd love to hear from you: info@learnyourenglish.com For more info on what we do at LYE, check out: Our Teacher Development Membership Join Our Mailing List Our Online CPD Courses for Teachers Follow Learn YOUR English Follow Teacher Talking Time
Since reinforcement learning requires hefty compute resources, it can be tough to keep up without a serious budget of your own. Find out how the team at Facebook AI Research (FAIR) is looking to increase access and level the playing field with the help of NetHack, an archaic rogue-like video game from the late 80s. Links discussed: The NetHack Learning Environment: https://ai.facebook.com/blog/nethack-learning-environment-to-advance-deep-reinforcement-learning/ Reinforcement learning, intrinsic motivation: https://arxiv.org/abs/2002.12292 Knowledge transfer: https://arxiv.org/abs/1910.08210 Tim Rocktäschel is a Research Scientist at Facebook AI Research (FAIR) London and a Lecturer in the Department of Computer Science at University College London (UCL). At UCL, he is a member of the UCL Centre for Artificial Intelligence and the UCL Natural Language Processing group. Prior to that, he was a Postdoctoral Researcher in the Whiteson Research Lab, a Stipendiary Lecturer in Computer Science at Hertford College, and a Junior Research Fellow in Computer Science at Jesus College, at the University of Oxford. https://twitter.com/_rockt Heinrich Kuttler is an AI and machine learning researcher at Facebook AI Research (FAIR) and before that was a research engineer and team lead at DeepMind. https://twitter.com/HeinrichKuttler https://www.linkedin.com/in/heinrich-kuttler/ Topics covered: 0:00 a lack of reproducibility in RL 1:05 What is NetHack and how did the idea come to be? 5:46 RL in Go vs NetHack 11:04 performance of vanilla agents, what do you optimize for 18:36 transferring domain knowledge, source diving 22:27 human vs machines intrinsic learning 28:19 ICLR paper - exploration and RL strategies 35:48 the future of reinforcement learning 43:18 going from supervised to reinforcement learning 45:07 reproducibility in RL 50:05 most underrated aspect of ML, biggest challenges? Get our podcast on these other platforms: Apple Podcasts: http://wandb.me/apple-podcasts Spotify: http://wandb.me/spotify Google: http://wandb.me/google-podcasts YouTube: http://wandb.me/youtube Soundcloud: http://wandb.me/soundcloud Tune in to our bi-weekly virtual salon and listen to industry leaders and researchers in machine learning share their research: http://wandb.me/salon Join our community of ML practitioners where we host AMA's, share interesting projects and meet other people working in Deep Learning: http://wandb.me/slack Our gallery features curated machine learning reports by researchers exploring deep learning techniques, Kagglers showcasing winning models, and industry leaders sharing best practices: https://wandb.ai/gallery
How Rosanne is working to democratize AI research and improve diversity and fairness in the field through starting a non-profit after being a founding member of Uber AI Labs, doing lots of amazing research, and publishing papers at top conferences. Rosanne is a machine learning researcher, and co-founder of ML Collective, a nonprofit organization for open collaboration and mentorship. Before that, she was a founding member of Uber AI. She has published research at NeurIPS, ICLR, ICML, Science, and other top venues. While at school she used neural networks to help discover novel materials and to optimize fuel efficiency in hybrid vehicles. ML Collective: http://mlcollective.org/ Controlling Text Generation with Plug and Play Language Models: https://eng.uber.com/pplm/ LCA: Loss Change Allocation for Neural Network Training: https://eng.uber.com/research/lca-loss-change-allocation-for-neural-network-training/ Topics covered 0:00 Sneak peek, Intro 1:53 The origin of ML Collective 5:31 Why a non-profit and who is MLC for? 14:30 LCA, Loss Change Allocation 18:20 Running an org, research vs admin work 20:10 Advice for people trying to get published 24:15 on reading papers and Intrinsic Dimension paper 36:25 NeurIPS - Open Collaboration 40:20 What is your reward function? 44:44 Underrated aspect of ML 47:22 How to get involved with MLC Get our podcast on these other platforms: Apple Podcasts: http://wandb.me/apple-podcasts Spotify: http://wandb.me/spotify Google: http://wandb.me/google-podcasts YouTube: http://wandb.me/youtube Tune in to our bi-weekly virtual salon and listen to industry leaders and researchers in machine learning share their research: http://wandb.me/salon Join our community of ML practitioners where we host AMA's, share interesting projects and meet other people working in Deep Learning: http://wandb.me/slack Our gallery features curated machine learning reports by researchers exploring deep learning techniques, Kagglers showcasing winning models, and industry leaders sharing best practices: https://wandb.ai/gallery
A new Episode of the Serie "The AI Deal of Trust" in the unique AI Chanel of Trust by AI Exponential Thinker. Our Guest is Dr. Anima Anandkumar is a Bren professor at Caltech and a director of machine learning research at NVIDIA. She was recognized by her peers with more than 20 awards from young investigator from the Air Force and Army research offices to Good Tech 2018 by Nytimes. She is one of youngest Female AI researcher in Silicon Valley with great Faculty fellowships from Microsoft, Google, Facebook and incredible best paper awards. Dr. Karoui is pleased to welcome Dr. Anandkumar in this new podcast episode. Dr. Lobna Karoui is an Executive AI Strategy Growth Advisor and Exponential Digital Transformer for Fortune 500 & CAC40 with two decades experience in building AI products and services for millions of users. She is the president of AI Exponential thinker with a target to inspire and empower 1 Million young boys and girls, horizon 2025, about Trust Technologies and AI Opportunities. She is an international Speaker and interviewer recognized as an AI Expert by Forbes, Bloomberg and MIT. Follow us and subscribe www.aiexponentialthinker.com, linkedin, Facebook and Instagram or via contact@aiexponentialthinker.com to interact with our Guests, meet great speakers and mentors from great companies such as Amazon, WEF, Harvard and more
Siddha is a self-driving architect at NVIDIA and she also guides teams at NASA as an AI domain expert. She was featured on the Forbes 30 under 30 list in 2019. Previously, she developed deep learning models for resource-constrained edge devices at DeepVision. She earned her Master's degree from Carnegie Mellon University, and her work ranges from visual question answering to generative adversarial networks to gathering insights from CERN’s petabyte-scale data and has been published at top-tier conferences including CVPR and NeurIPS.Hosted by:Jay Shah: https://www.linkedin.com/in/shahjay22/You can reach out to https://www.public.asu.edu/~jgshah1/ for any queries.Stay tuned for upcoming webinars!***Disclaimer: The information contained in this video represents the views and opinions of the speaker and does not necessarily represent the views or opinions of any institution. It does not constitute an endorsement by any Institution or its affiliates of such video content.***
Professor Julie Dockrell speaks to Dr Rob Webster about how speech, language and communication are foundational to learning and achievement. What happens when a child struggles with these skills? Slow progress can be hidden by the logistics and pressures of the current school system, making oral language delay a less well-recognised disorder, which is in turn less supported as children move up through school years. Disorders such as development language disorder (DLD) represent the most extreme end of difficulties. Rob and Julie discuss reasons why some children might be slower than others to develop skills in this arena, and how they can be supported to overcome difficulties. We also hear examples of techniques that teachers can employ to create more opportunities for oral language development during classroom activities. Full show notes and links: https://www.ucl.ac.uk/ioe/news/2020/oct/overcoming-oral-language-barriers-learning-rftrw-s05e03 Take our 2-minute survey to share your thoughts about our podcast: https://bit.ly/ResearchPodcast
Dr. Megan Sumeracki teaches us about Spacing, Retrieval, Interleaving, Elaboration, Dual-coding, and Concrete Examples. [2:16] The Learning Scientists Podcast & Bi-directional Communication [4:13] The Importance of Learning Research as open-access [5:27] The Six Strategies of Effective Learning from the Institute for Education Sciences (2007?) & The National Council for Teaching Quality Report (2016?) [7:45] A Brief History of Spacing and Retrieval Practice, Ebbinghaus, and Memory Accessibility [9:57] What is Spacing and How do We Use it? [12:17] The Ins and Outs of Retrieval Practice [4:47] Interleaving Your Study Practice (19:10 Taylor and Rore? 2010 blocking v. interleaving) [21:20] Elaborative Interrogation: The How, When, Why, and Where of Your Study Material [24:40] Dual Coding: Combining Visual information with your Learning! [28:05] Concrete Examples: The More the Merrier [32:54] Closing Advice from Dr. Sumeracki: Use These Strategies!!! Also, you can find Dr. Sumeracki’s books on Amazon, research articles, and some great podcasts by the Learning Scientists on Retrieval for medical residents, and an interview with her sister, Dr. Alyssa Smith.
What we know about this coming school year, is how much we don’t know what will happen or what the school year is going to end up looking like. Kids returning to school usually has us feeling mixed emotions. We are excited to have them back to school, back to routine but we also can feel melancholy at the site of our kids entering a new grade, getting older and more independent. Now we need to add in the uncertainty of going back to school during a pandemic. How are you feeling about it? Are you nervous? Worried about your child not being safe? Worried school might be closed again (and terrified of having to homeschool your kids again!)? Worried that they are falling behind in their education? Worried about their future? I get it. My guest, Dr. Annie Snyder and I talk about all these fears and worries in this episode and what to do about it for yourself, teachers, and for your kids. Dr. Snyder is an educational researcher who is studying the “Covid Slide” which is the decrease in learning due to the lengthy disruption of school from last year plus the summer holidays. This episode is full of useful tips and advice on how to handle returning to school. Be sure to check out the show notes for important tips from the episode but also for Dr. Snyder’s tips on how to support your kids with continual learning. About Dr. Annie Snyder A former teacher and educational researcher, Annie Snyder currently serves as a senior learning scientist within the Learning Research and Strategy team at McGraw Hill. She holds a doctorate in educational psychology from Teachers College, Columbia University, and is interested in the links between learning science, research, teaching practice, and families’ roles in education. Annie has been happily working from home with her three sons since 2016. From the Podcast Dr. Snyder has so many great resources to keep the learning going and on how to support your kids during this time. Tips from the Podcast: Encourage young learners to exercise for a few minutes before any online learning (jumping jacks, dancing, running in place, etc.). Physical movement will help those learners become more ready to engage in learning and stay focused. Make a written checklist for any procedures that learners may need to follow, whether they are going to school online or in a brick and mortar school (for example, for online learning, turning on the computer, shutting the door, drinking some water, preparing materials before logging on). Especially for new processes, this can help prevent learners from needing to use their cognitive resources on procedures, so they can more easily pay attention to school learning. Explore the accessibility options offered through technology (e.g. text-to-speech options for online text). Some of these may help ease learning burdens (and boredom!) by providing learners other avenues for accessing learning content. Try forming online (or safe, socially distant) study and social groups that meet on a regular basis. Learners can not only have the opportunity both to work on school tasks, but this also offers an opportunity for practicing social skills as well as working through some of the emotional challenges posed by learning during a pandemic. As best as possible, reduce distractions when learners must learn online. These distractions might be in the room itself (e.g. move the toys out of the room BEFORE the virtual meeting with the class) but also distractions within the technology (so that learners are not surfing the Internet when they need to be responding to a teacher’s question). Work together, as a family and with educators. Continuously remind yourself and those around you that we will succeed by being kind, empathetic, and cooperative to each other as we all work toward ensuring children make forward progress. Remember that learning never stops. The human brain is continuously learning, all throughout the lifespan, and the world is rich with opportunities to continue to grow and learn more – no matter where we all might be in any given moment. Dr. Snyder’s Tips to Keep Learning Going Literacy: Read aloud: All forms of reading aloud can support literacy learning for children of all ages. While it is always terrific when adults can read aloud to children, remember that children can also read to adults, other children, pets, or stuffed animals. Children can also benefit from listening to online read-alouds, as well as recording their own. Lean on the library: Throughout the country, libraries have been hard at work finding new ways to provide services online. Check out the local library to determine whether there are online programs, summer reading programs, eBook loans, book recommendations and other offerings to support literacy learning this summer. Encourage writing: Some children may naturally want to engage in writing as a way to express their ideas and emotions, while others may need more encouragement. Families can support writing development by setting aside designated writing periods each day and providing a few basic materials such as writing prompts, blank comic book strips, journals, or even stamps and envelopes for written correspondence. Remember that writing doesn’t necessarily mean that children must put pen to paper – typing and oral recordings also count! Math: Play games: A deck of cards, a sheet of paper and a pencil, and a few dice can all provide countless opportunities for playing math games (for example, Addition “War” with cards, in which each player plays two cards instead of one and the player with the highest sum wins the cards). Chess, checkers, and other classic board games are also great choices, as they can support the abstract strategic thinking that is critical for mathematical development. Discover real-world math: Math surrounds us at every turn, and it can be exciting for children (and their families!) to uncover all the ways math can not only help us understand the world, but also solve problems. Whether children are measuring for a garden, tracking weather statistics, or creating a budget to help save for a new toy, authentic projects help younger learners make sense of mathematics. Make math active: Try skip count hopscotch with sidewalk chalk (count by tens for each square!), jump rope counting challenges, or number line dancing. All of these are just a few ways that families can blend mathematical learning with movement. This is especially helpful for younger children, who may be less inclined to want to sit and practice with flashcards or worksheets during enticing summer weather. Science: Join a community: Although we often think of lone scientists working in a lab when we hear the word science, in reality, many scientists work collaboratively as part of scientific communities. Young learners can be encouraged to form their own family science communities (e.g. a backyard experiment family club) or can, with adult help, take part in one of the many citizen science projects that are available online. Explore: Both the natural and the human-made environments that surround us can provide lots of options for scientific exploration. For example, nature hikes or city walks following appropriate social distancing precautions can provide families with opportunities to help children explore science in the world around them (e.g. birdwatching, tree identification, studies of architectural principles of city buildings, etc.) Engineer: Summer is an excellent time for building things with blocks, cardboard boxes, or whatever scrap materials happen to be on hand. For younger children and older children alike, it can be helpful to provide daily challenges (e.g. build the tallest tower, the longest tunnel, a maze, etc.) to help inspire problem-solving and collaborative engineering. Social Studies: Study civics: What is civics, exactly? This single question can open doors to endless writing opportunities, discussions, and projects during the summer months. In addition to examining definitions of civics, learners and their families can explore ways to engage in civics both at the family level (e.g. creating a list of family rules for the summer) as well as at other levels of society (e.g. learning how the various branches of government work together, studying the Constitution, writing letters to Congress, reading the news). Build an exhibit: Museums around the world are offering virtual tours of their exhibits. After taking a tour, learners can learn more about some of the artifacts and then create an exhibit (or an entire museum) of their own! Cardboard boxes, an empty shelf, a table, or a blanket on the floor could be a great place to display the exhibit. To add authenticity, learners may wish to make placards to provide facts for every artifact included. Older learners might want to take digital photos of the artifacts and record a gallery walk for virtual visitors. Travel the world, from home: The study of geography is all about the places, physical features, and people around the world - and how all these interact. Although actual travel may not be a possibility for families this summer, learning about the world can happen from anywhere. For example, families can explore atlases (online and print), practice map skills by creating maps of the home or neighborhood or investigate the geographic origins of objects in the home. Thanks for listening! It means so much to me that you listened to my podcast! If you would like to purchase my book or other parenting resources, visit me at www.yellingcurebook.com With this podcast, my intention is to build a community of parents that can have open and honest conversations about parenting without judgement or criticism. We have too much of that! I honor each parent and their path towards becoming the best parent they can be. My hope is to inspire more parents to consider the practice of Peaceful Parenting. If you know somebody who would benefit from this message, or would be an awesome addition to our community, please share it using the social media buttons on this page. Do you have some feedback or questions about this episode? Leave a note in the comment section below! Subscribe to the podcast If you would like to get automatic updates of new podcast episodes, you can subscribe on the podcast app on your mobile device. Leave a review I appreciate every bit of feedback to make this a value adding part of your day. Ratings and reviews from listeners not only help me improve, but also help others find me in their podcast app. If you have a minute, an honest review on iTunes goes a long way! Thank You!!
Today, we are joined by Dr. Louis Gomez, senior fellow at the Carnegie Foundation for the Advancement of Teaching and professor of education and information studies at UCLA. Dr. Gomez addresses the topics of Networked Improvement Communities (NICs) and the mindset shift that is necessary in order to promote equity in education. Join our conversation about how compliance can prevent initiatives from being implemented, the necessity of having a common aim and narrative when discussing improvement science as part of NICs, and that equity without the will to change or the respect for the community will not bring about social justice. Visit our Bronx ART website and connect with us on Twitter @BX_ARTeam! Today's hosts are Kris DeFilippis, Adelia Gibson, and Kaitlyn Reilley Guest Information: Dr. Gomez earned a bachelor's degree in psychology from Stonybrook University and a doctorate in Psychology from University of California, Berkeley. He spent 14 years working in cognitive science and person–computer systems interactions at Bell Laboratories, Bell Communications Research Inc. and Bellcore. Dr. Gomez has held a number of faculty positions including positions at Northwestern University and the University of Pittsburgh, where he was also director of the Center for Urban Education and a senior scientist at the Learning Research and Development Center. Dr. Gomez is currently a professor of education and information studies at the University of California, Los Angeles. Since 2008, he has served as a senior fellow at the Carnegie Foundation for the Advancement of Teaching, where he leads the Network Development work. He is the co-author of Learning to Improve: How America's Schools Can Get Better at Getting Better. Dr. Gomez is dedicated to educational improvement and his numerous publications and studies have contributed greatly to bringing improvement science to the field of education. Connect with Dr. Gomez through email at lmgomez@ucla.edu Resources for Listeners: Information about iLEAD Learning to Improve: How America's Schools Can Get Better and Getting Better Why a NIC? Getting Ideas into Action: Building Networked Improvement Communities in Education Improvement Research Carried Out Through Networked Communities: Accelerating Learning about Practices that Support More Productive Student Mindsets How a Networked Improvement Community Improved Success Rates for Struggling College Math Students
In the seventh episode of Deep Neural Notebooks, I interview Shimon Whiteson. Shimon sir is a Computer Science Professor at the University of Oxford, where he leads the Whiteson Research Lab. He is also a Data Scientist at Waymo (formerly the Google Self Driving Car Project). His research specialises in Reinforcement Learning (RL), Cooperative Multi-Agent RL, to be precise. So this interview is all in the context of Reinforcement Learning. We talk about his journey - how he started with Machine Learning & RL. I ask him about his thoughts on the state of RL - about how the field has progressed and changed since he started, about how it has become so popular in the last few years, and about the challenges being faced. We also talk about his research at Waymo, about recent projects from his lab, and about the scope and future of telepresence robots, one of which was developed under his guidance. We also talk about the infamous Reward Hypothesis in the context of RL and Philosophy. In the end, he also shares some advice for people starting out with RL. Links: - Shimon Whiteson: https://twitter.com/shimon8282 - Whiteson Research Lab (WhiRL): http://whirl.cs.ox.ac.uk/ - Teresa Robot: https://whirl.cs.ox.ac.uk/teresa/ - RL workshop at Machine Learning Summer School, Moscow: https://www.youtube.com/watch?v=RAw0Chs7QKA - The Reward Hypothesis: http://incompleteideas.net/rlai.cs.ualberta.ca/RLAI/rewardhypothesis.html Timestamps: 03:42 Beginnings in Computer Science06:13 Beginnings in ML 07:15 PhD at UT Austin 10:40 Intersection of Neuroevolution and RL 14:10 Research directions since PhD 16:35 State of RL 20:33 Simulation for RL 22:07 Research at Waymo 25:30 Multi-agent RL 33:25 Recent projects at WhiRL 41:30 Teresa project and Telepresence Robots 48:08 Bottlenecks for RL and Robotics 49:45 End-goal for RL, Human-level Intelligence 53:45 What do you find most fascinating about your research? 55:38 RL & Philosophy 1:01:20 Keeping up with latest research 1:03:28 Advice for beginners Podcast links : Youtube: https://youtu.be/bbrYZDgPI9M Apple Podcasts: https://apple.co/2TLUZ0y Google Podcasts: https://bit.ly/2TIyvh6 Spotify: https://open.spotify.com/episode/3936aEvSwsIhfwQfURmDb9 Anchor: https://bit.ly/3gpMi65 Connect: Twitter: https://twitter.com/mkulkhanna Website: https://mukulkhanna.co LinkedIn: https://linkedin.com/in/mukulkhanna/
Our guest Malte Pietsch is a Co-Founder of deepset, where he builds NLP solutions for enterprise clients, such as Siemens, Airbus and Springer Nature. He holds a M.Sc. with honors from TU Munich and conducted research at Carnegie Mellon University.He is an active open-source contributor, creator of the NLP frameworks FARM & haystack and published the German BERT model. He is particularly interested in transfer learning and its application to question answering / semantic search.Resources:Deepset - Make sense out of your text data - https://deepset.ai/FARM - Fast & easy transfer learning for NLP - https://github.com/deepset-ai/FARMHayStack - Transformers at scale for question answering & search - https://github.com/deepset-ai/haystackSageMaker - Machine learning for every developer and data scientist - https://aws.amazon.com/sagemaker/Spot Instances - Managed Spot Training in Amazon SageMaker - https://docs.aws.amazon.com/sagemaker/latest/dg/model-managed-spot-training.htmlElasticSearch - Fully managed, scalable, and secure Elasticsearch service - https://aws.amazon.com/elasticsearch-service/Automatic mixed precision - Automatic Mixed Precision for Deep Learning - https://developer.nvidia.com/automatic-mixed-precisionPyTorch - Open source machine learning framework that accelerates the path from research prototyping to production deployment - https://pytorch.org/NumPy - Fundamental package for scientific computing with Python - https://numpy.org/MLFlow - An open source platform for the machine learning lifecycle - https://mlflow.org/BERT - Bidirectional Encoder Representations from Transformers - https://en.wikipedia.org/wiki/BERT_(language_model)SQuAD - The Stanford Question Answering Dataset - https://rajpurkar.github.io/SQuAD-explorer/Sebastian Ruder - Research scientist at DeepMind - https://ruder.io/Andrew Ng - His machine learning course is the MOOC that had led to the founding of Coursera - https://www.coursera.org/instructor/andrewng
On Ep 16 of BITS Cast, I have Varuni Sarwal on the show. She is an undergraduate at IIT Delhi and has interned at Top Colleges like Harvard University and UCLA, for Deep Learning. She is passionate about the application of Deep Learning in Biology. She shares some of her tips for how you too can get an international Internship. She goes into detail and shares her experience at these colleges and what all she learned and what are her future plans for this new decade. Let me know what was your biggest takeaway on Instagram @ishansharma7390
Following up on our recent show covering the educational research highlights of 2019, Mike sits down with Youki Terada from Edutopia who authored the article. Youki is the Research and Standards editor for Edutopia which means he reviews and edits contributions from Edutopia's writing staff to ensure it's evidence-based, well-designed, and relevant to Edutopia's target audience of K12 Educators. We talk about areas of research that Youki has found particularly interesting and explore several examples with an eye towards practical application for educators. We also talk about the importance of curation and the risks of fast or sloppy research when looking for good applications of emerging learning research.
Dan and Mike dig into an Edutopia article by Youki Terada outlining the key findings in educational research in 2019. What sorts of findings jump out and which stories did we cover on Trending in Education? We take some time to look back as we gear up to peer forward into 2020 and the decade to come on this week’s episode. We hope you enjoy!
Personal Note: I'm extremely excited and really honoured to share this interview! Show notes: https://sanyambhutani.com/interview-with-jeremy-howard/ Audio (Podcast Version) available here: https://anchor.fm/chaitimedatascience Subscribe here to the newsletter: https://tinyletter.com/sanyambhutani In this episode, Sanyam Bhutani interviews Jeremy Howard: Co-Founder at fast.ai and Guru to the complete fast.ai community. They talk all about Jeremy's journey into through the field as a Consultant to Founder to Kaggle Rank 1 followed by founding fast.ai. They also discuss all about the fast.ai forums: The research that goes into it, how Jeremy and Sylvain collaborate on research and how the course is prepared. Links: fast.ai course: https://course.fast.ai Interview with Sylvain Gugger: Interview with Jason Antic (Creator of DeOldify): Follow: Jeremy Howard: https://twitter.com/jeremyphoward Fast.ai: https://twitter.com/fastdotai https://fast.ai Sanyam Bhutani: https://twitter.com/bhutanisanyam1 Blog: sanyambhutani.com About: http://chaitimedatascience.com/ A show for Interviews with Practitioners, Kagglers & Researchers and all things Data Science hosted by Sanyam Bhutani. You can expect weekly episodes every available as Video, Podcast, and blogposts. If you'd like to support the podcast: https://www.patreon.com/chaitimedatascience Intro track: Flow by LiQWYD https://soundcloud.com/liqwyd --- Send in a voice message: https://anchor.fm/chaitimedatascience/message
Newsletter: https://tinyletter.com/sanyambhutani Personal Note: I'm really honored to share this conversation. I really hope you enjoy listening to it as much as I enjoyed talking to Dr. Marc Lanctot. In this Episode, Sanyam Bhutani interviews Dr. Marc Lactot, a Research Scientist at DeepMind. In this interview, they talk all about Research at DeepMind, Deep Learning Research, AlphaGo. They also talk all about Swift For Tensorflow and OpenSpiel. Dr. Marc Lanctot is a research scientist at Google DeepMind. Previously, he was a post-doctoral researcher at the Maastricht University Games and AI Group, working with Mark Winands. During his PhD, Marc worked on sampling algorithms for equilibrium computation and decision-making in games. His research interests include in general multiagent learning (and planning), computational game theory, reinforcement learning, and game-tree search. Open Spiel: https://deepmind.com/research/open-source/openspiel Interview with Machine Learning Hero Series: https://medium.com/dsnet/interviews-with-machine-learning-heroes-ad9358385278 Links from the podcast: - Commodore 64: https://en.wikipedia.org/wiki/Commodore_64 - Intellivision (early console!): https://en.wikipedia.org/wiki/Intellivision - AD&D Treasure of Tarmin (Intellivision game with 3D perspective-- pretty advanced for its time!): https://en.wikipedia.org/wiki/Advanced_Dungeons_%26_Dragons:_Treasure_of_Tarmin - Montezuma's revenge: https://en.wikipedia.org/wiki/Montezuma%27s_Revenge_(video_game) Follow: Dr. Marc Lanctot https://twitter.com/sharky6000 http://mlanctot.info/ https://www.linkedin.com/in/marc-lanctot-a1619095/?originalSubdomain=ca Sanyam Bhutani: https://twitter.com/bhutanisanyam1 About: http://chaitimedatascience.com/ A show for Interviews with Practitioners, Kagglers & Researchers and all things Data Science hosted by Sanyam Bhutani. You can expect weekly episodes every Sunday, Thursday available as Video, Podcast, and blogposts. If you'd like to support the podcast: https://www.patreon.com/chaitimedatascience --- Send in a voice message: https://anchor.fm/chaitimedatascience/message
Machine Learning merupakan sebuah algoritma yang digunakan untuk mempelajari sebuah data, tentang bagaimana pola, dan membuat model berdasarkan data historik. Data yang nanti dihasilkan berguna sebagai referensi, klasifikasi dan prediksi. Hendri Krisma (Senior Development Engineer Blibli) menjelaskan bagaimana Machine Learning tersebut dilakukan dengan teknologi berbasis JVM.
Audio (Podcast Version) available here: https://anchor.fm/chaitimedatascience Subscribe here to the newsletter: https://tinyletter.com/sanyambhutani In this Episode, Sanyam Bhutani interviews Pierre Stock, resident PhD student at FAIR Lab, Paris. In this interview, they talk a lot about his research, Deep Learning Research in general and FAIR. They also discuss Pierre's recent research: And the Bit Goes Down: Revisiting the Quantization of Neural Networks (Link: https://arxiv.org/pdf/1907.05686.pdf) Repository: https://github.com/facebookresearch/kill-the-bits Follow: Pierre Stock: https://twitter.com/PierreStock https://research.fb.com/people/stock-pierre/ Sanyam Bhutani: https://twitter.com/bhutanisanyam1 Blog: https://medium.com/@init_27 About: http://chaitimedatascience.com/ A show for Interviews with Practitioners, Kagglers & Researchers and all things Data Science hosted by Sanyam Bhutani. You can expect weekly episodes every available as Video, Podcast, and blogposts. If you'd like to support the podcast: https://www.patreon.com/chaitimedatascience --- Send in a voice message: https://anchor.fm/chaitimedatascience/message
Previous Blog Interview: https://medium.com/dsnet/interview-with-openai-fellow-christine-mcleavey-payne-aaef948ad571 In this episode, Sanyam Bhutani interviews Christine Payne for the second time. They talk about MuseNet, OpenAI and Deep Learning Research, MOOC(s) for Machine Learning, Deep Learning. MuseNet: https://openai.com/blog/musenet/ Follow: Christine Payne: https://twitter.com/mcleavey https://www.linkedin.com/in/mcleavey/ http://christinemcleavey.com/ Sanyam Bhutani: https://twitter.com/bhutanisanyam1 About: http://chaitimedatascience.com/ A show for Interviews with Practitioners, Kagglers & Researchers and all things Data Science hosted by Sanyam Bhutani. You can expect weekly episodes every available as Video, Podcast, and blogposts. If you'd like to support the podcast: https://www.patreon.com/chaitimedatascience --- Send in a voice message: https://anchor.fm/chaitimedatascience/message
This has been one of the most enlightening conversations for us about Deep Learning Research. In this Interview, Sanyam Bhutani interviews Tim Dettmers, a Ph.D. student at the University of Washington. They talk about Deep Learning Research, Pipelines for Research, Sparse Networks from Scratch: Faster training without Losing Performance paper by Tim. They also talk about Hardware, Tim's Kaggle. Follow: Tim Dettmers: Twitter, Blog Sanyam Bhutani: Twitter, Blog About: http://chaitimedatascience.com/ A show for Interviews with Practitioners, Kagglers & Researchers and all things Data Science hosted by Sanyam Bhutani. You can expect weekly episodes available as Video, Podcast, and blogposts. If you'd like to support the podcast: https://www.patreon.com/chaitimedatascience --- Send in a voice message: https://anchor.fm/chaitimedatascience/message
I caught up with Hilary Mason, GM of Machine Learning at Cloudera and former founder of Fast Forward Labs. We cover how to: - Generate ideas for Machine Learning Research - Hold a good brainstorm - Break into tech as a Data Scientist - Give an engaging and successful tech talk I also asked Hilary about the evolving role of Machine Learning and Data Science, and what she wants to build in the future!
Get my 5 Tips To Address Implicit Bias Within Ourselves and Others About Lindsay Page, Ph.D. Lindsay Page is an assistant professor of research methodology at the University of Pittsburgh School of Education, a research scientist at Pitt’s Learning Research and Development Center, and a faculty research fellow of the National Bureau of Economic Research. Her work focuses on quantitative methods and their application to questions regarding the effectiveness of educational policies and programs across the pre-school to postsecondary spectrum. Much of her recent work has involved the implementation of large-scale randomized trials to investigate innovative strategies for improving students’ transition to and through college. She holds a doctorate in quantitative policy analysis and master's degrees in statistics and in education policy from Harvard University. Show Highlights A pilot program to combat chronic absenteeism Using a text-messaging service to communicate with families Organization and logistics of the text-messaging communication Multiple benefits to the two-way texting system Results and feedback from the pilot program Sending communication for multi-lingual families The impact of the pilot program on chronic absenteeism Establishing the definition of chronic absenteeism Getting a two-way text-messaging service established Connect with Lindsay Twitter: @linzcpage Email: lpage@pitt.edu Additional Resources Connect-Text: Leveraging Text-Message Communication to Mitigate Chronic Absenteeism and Improve Parental Engagement in the Earliest Years of Schooling Connect with me on Twitter @sheldoneakins www.sheldoneakins.com
Mrs. Gilpin gave a presentation at the Virginia State Reading Association (VSRA) annual conference in Norfolk, Va. in March. Her presentation revolved around her recently authored book; Teaching Students to Conduct Short Research Projects. Listen in to the podcast above as Mrs. Gilpin walks you through the five major steps in the research process with examples from her own classroom experience. Mrs. Gilpin would like to thank former Powhatan teacher and national consultant, Mrs. Laura Robb, for her amazing help and guidance with this endeavor. Read more: http://powhatanschool.org/pulse/food-for-thought/gilpin-lifelong-learning-research-book/
Dr. Megan Sumeracki teaches us about Spacing, Retrieval, Interleaving, Elaboration, Dual-coding, and Concrete Examples. You will not want to miss this episode covering the evidence-based learning strategies that are proven to be the most effecting study methods. On today’s show, Dr. Megan Sumeracki will teach us about Spacing, Retrieval, Interleaving, Elaboration, Dual-coding, and Concrete Examples. Using these methods are a proven way to make your study time more efficient. Intro 2:16 The Learning Scientists Podcast & Bi-directional Communication 4:13 The Importance of Learning Research as open-access 5:27 The Six Strategies of Effective Learning from the Institute for Education Sciences (2007?) & The National Council for Teaching Quality Report (2016?) 7:45 A Brief History of Spacing and Retrieval Practice, Ebbinghaus, and Memory Accessibility 9:57 What is Spacing and How do We Use it? 12:17 The Ins and Outs of Retrieval Practice 14:47 Interleaving Your Study Practice (19:10 Taylor and Rore? 2010 blocking v. interleaving) 21:20 Elaborative Interrogation: The How, When, Why, and Where of Your Study Material 24:40 Dual Coding: Combining Visual information with your Learning! 28:05 Concrete Examples: The More the Merrier 32:54 Closing Advise from Dr. Sumeracki: Use These Strategies!!! Also, you can find Dr. Sumeracki’s books on Amazon, research articles, and some great podcasts by the Learning Scientists on Retrieval for medical residents, and an interview with her sister, Dr. Alyssa Smith.
Thank you to Cool Initiatives for sponsoring this week's episode. Check out their competition, and chance to win £10k, here. What's in this episode? Hello everyone, this week we listen into the Secretary of State for Education's plans for Edtech in England, plus the creative brains behind Scratch on the early days of computers, a teacher and the institute of imagination on hands on learning, a new club for edutainment toys, and an insanely enthusiastic digital media FE teacher on collecting masters and delving into research on AI, our emotions and learning. Enjoy! People Sophie Bailey is the Founder and Presenter of The Edtech Podcast | Twitter: @podcastedtech Damian Hinds, Secretary of State for Education and MP for East Hampshire | Twitter: @DamianHinds Mitchel Resnick, Professor of Learning Research at MIT Media Lab, author of Lifelong Kindergarten, and founder of the Scratch Team | Twitter: @mres Shafina Vohra, Teacher of A Level Psychology at the London Design & Engineering UTC | Twitter: @ShafinaVohra Annie Duffield, Head of Marketing and Communications, Institute of Imagination | Twitter: @annieduffs Ben Callicot, Founder, Toy Pioneers Club Ian Hurd, Media & Creative Arts Teacher Show Notes and References Check out https://theedtechpodcast.com/edtechpodcast for the full show notes. Tell us your story We'd love to hear your thoughts. Record a quick free voicemail via speakpipe for inclusion in the next episode. Or you can post your thoughts or follow-on links via twitter @podcastedtech or via The Edtech Podcast Facebook page or Instagram.
The T4L team step aside to make way for student interviewers Declan and Bhagyada from Greystanes High School, as they spend the afternoon with one of the creators of Scratch Professor Mitchel Resnick! Professor Resnick is LEGO Papert Professor of Learning Research, Director of the Okawa Center, and Director of the Lifelong Kindergarten group at the MIT Media Lab. A conversation not to be missed!
Today, Melanie brings you another great interview from her time at Deep Learning Indaba in South Africa. She was joined by Yabebal Fantaye and Jessica Phalafala for an in-depth look at the deep learning research that’s going on in the continent. At the African Institute for Mathematical Sciences, the aim is to gather together minds from all over Africa and the world to not only learn but to use their distinct perspectives to contribute to research that furthers the sciences. Our guests are both part of this initiative, using their specialized skills to expand the abilities of the group and stretch the boundaries of machine learning, mathematics, and other sciences. Yabebal elaborates on the importance of AIMS and Deep Learning Indaba, noting that the more people can connect with each other, the more confidence they will gain. Jessica points out how this research in Africa can do more than just advance science. By focusing on African problems and solutions, machine learning research can help increase the GDP and economic standards of a continent thought to be “behind”. Jessica Phalafala Jessica Phalafala is a PhD Applied Mathematics student at Stellenbosch University and currently affiliated with the African Institute for Mathematical Sciences. In her mid-twenties, she finds herself with four qualifications all obtained with distinction, including a Master of Science in Pure Mathematics degree from the University of the Witwatersrand. Jessica is interested in using her functional analysis background together with a number of newly developed skills to contribute towards developing rigorous mathematical theory to support some existing deep learning methods and algorithms for her PhD research. Outside of research she takes great interest in fast-tracking the level of accessibility of higher education in South Africa as co-founder of the Sego Sa Lesedi Foundation, a platform created to inform underprivileged high school learners of career and funding opportunities in science as well as provide them with mentorship as they transition into undergraduate studies. Yabebal Fantaye Dr. Fantaye is an AIMS-ARETE Research Chair based in South Africa. His research is in applying artificial intelligence and advanced statistical methods to cosmological data sets in order to understand the nature of the universe and to satellite images of the Earth in order to find alternative ways to monitor African development progress. Dr. Fantaye is a fellow of the World Economic Forum Young Scientists community, and a fellow and a Chair of the Next Einstein Forum Community of Scientists. Cool things of the week A Kubernetes FAQ for the C-suite blog BigQuery and surrogate keys: a practical approach blog Adding custom intelligence to Gmail with serverless on GCP blog Announcing Cloud Tasks, a task queue service for App Engine flex and second generation runtimes blog Unity and DeepMind partner to advance AI research blog Interview African Institute for Mathematical Sciences site Provable approximation properties for deep neural networks research Next Einstein Initiative site Square Kilometer Array (SKA) site University of the Witwatersrand site Council of Scientific and Industrial Research (CSIR) site South African National Space Agency (SANSA) site National Astrophysics and Space Science Programme (NASSP)site IndabaX site Coursera site Andrej Karpathy research Andrej Karpathy Blog blog Question of the week If I’m using the Cluster Autoscaler for Kubernetes (or GKE), how can I prevent it from removing specific nodes from the cluster when scaling down? How can I prevent Cluster Autoscaler from scaling down a particular node? github What types of pods can prevent CA from removing a node? github Where can you find us next? Mark will definitely be at Kubecon in December and will probably be at Unite L.A. this month. Melanie is speaking at Monktoberfest Oct 4th in Portland, Maine and will be at CAMLIS the following week.
SuperCreativity Podcast with James Taylor | Creativity, Innovation and Inspiring Ideas
Mitch Resnick, Professor of Learning Research at the MIT Media Lab, develops new technologies and activities to engage children in creative learning experiences. He was centrally involved in the development of the Scratch programming language, the LEGO Mindstorms robotics kits, and the Computer Clubhouse network of after-school learning centers. He has been awarded with the […] The post CL187: Cultivating Creating In Children And Adults – Interview with Mitchel Resnick appeared first on James Taylor.
Established in 2008 as a partnership between the City of Pittsburgh, Pittsburgh Public Schools, and the University of Pittsburgh Medical Center, The Pittsburgh Promise has since provided more than $120 million in college scholarships to eligible graduates of Pittsburgh's public schools. But is it working? On this episode of Research Minutes, University of Pittsburgh School of Education researcher Lindsay Page speaks with CPRE researcher Robert Nathenson (University of Pennsylvania) about her recent study of the Pittsburgh Promise and its impacts on college enrollment and persistence. Her study, The Promise of Place-Based Investment in College Access and Success: Investigating the Impact of the Pittsburgh Promise, was co-authored by Jennifer Iriti, Danielle Lowry, and Aaron Anthony (University of Pittsburgh). It was published in Education Finance and Policy in 2018. Lindsay Page is an assistant professor of research methodology at the School of Education and a research scientist at the Learning Research and Development Center at the University of Pittsburgh. Her work focuses on quantitative methods and their application to questions regarding the effectiveness of educational policies and programs across the pre-school to postsecondary spectrum. --- To learn more about The University of Pennsylvania's CPRE Knowledge Hub, visit http://www.cprehub.org/ or follow @cprehub.
Bryan Catanzaro, the VP Applied Deep Learning Research at NVIDIA, joins Mark and Melanie this week to discuss how his team uses applied deep learning to make NVIDIA products and processes better. We talk about parallel processing and compute with GPUs as well as his team's research in graphics, text and audio to change how these forms of communication are created and rendered by using deep learning. This week we are also joined by a special co-host, Sherol Chen who is a developer advocate on GCP and machine learning researcher on Magenta at Google. Listen at the end of the podcast where Mark and Sherol chat about all things GDC. Bryan Catanzaro Bryan Catanzaro is VP of Applied Deep Learning Research at NVIDIA, where he leads a team solving problems in domains ranging from video games to chip design using deep learning. Bryan earned his PhD from Berkeley, where he focused on parallel computing, machine learning, and programming models. He earned his MS and BS from Brigham Young University, where he worked on higher radix floating-point representations for FPGAs. Bryan worked at Baidu to create next generation systems for training and deploying deep learning models for speech recognition. Before that, he was a researcher at NVIDIA, where he worked on programming models for parallel processors, as well as libraries for deep learning, which culminated in the creation of the widely used CUDNN library. Cool things of the week NVIDIA Tesla V100s coming to Google Cloud site Automatic Serverless Deployment with Cloud Source Repositories blog Magenta site NSynth Super site MusicVAE site Making music using new sounds generated with machine learnnig blog Building Blocks of Interpretability blog Interview NVIDIA site NVIDIA GPU Technology Conference (GTC) site CUDA site cuDNN site NVIDIA Volta site NVIDIA Tesla P4 docs NVIDIA Tesla V100s site Silicon Valley AI Lab Baidu Research site ICML: International Conference on Machine Learning site CVPR: Computer Vision and Pattern Recognition Conference site Referenced Papers & Research: Deep learning with COTS HPC System paper Building High-level Features Using Large Scale Unsupervised Learning paper OpenAI Learning to Generate Reviews and Discovering Sentiment paper Progressive Growing of GANs for Improved Quality, Stability, and Variation paper and CelebA dataset High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs paper Deep Image Prior site How a Japanese cucumber farmer is using deep learning and TensorFlow blog Sample Talks: Future of AI Hardware Panel video High Performance Computing is Supercharging AI blog/video AI Podcast: Where is Deep Learning Going Next? blog/video Sample Resources: Coursera How Google does Machine Learning site NVIDIA Deep Learning Institute site Udacity AI Nanodegree site Kaggle site TensorFlow site PyTorch site Keras site Question of the week What to watch out for and get involved in at the Game Developers Conference (GDC) this year and in the future? International Grame Developers Association (IGDA) site Fellowship of GDC Parties site ALtCtrlGDC site Experimental Gameplay Workshop site Women in Games International (WIGI) site Blacks in Gaming (BIG) site Serious Games (SIGs) site What's New in Firebase and Google Cloud Platform for Games site Summits to Checkout: AI Game Developers Summit site Game Narrative Summit site Independent Games Summit site Additional Advice: The first two days are summits which are great because topic focused Expo floor takes a good hour to get through WIGI, BIG and SIGs (Google and Microsoft) have the best food GDC is composed of various communities Bring business cards Check out post-mortems Favorite Games: Mass Effect site Final Fantasy wiki Games Mark & Sherol are currently playing: Hearthstone site Dragon Age Origins wiki Where can you find us next? Mark and Sherol are at the Game Developer's Conference (GDC). You can find them via the Google at GDC 2018 site. Sherol will be at TensorFlow Dev Summit speaking about machine learning research and creativity next week.
Podcart — People need fresh, healthy food. But growing fresh food needs a reliable source of water in a water scarce country. The Amanzi for Food project – funded by the Water Research Commission (WRC) and led and implemented by the Environmental Learning Research Centre at Rhodes – brings together farmers, extension services, local economic development, agricultural training institutions and agricultural NGOs. They share knowledge and skills around harvesting, storing, and using rainwater to improve food production and to make farming as sustainable as possible. Find out how it works in this first edition of ‘Engagement in Action’ and how it benefits every single person involved – from subsistence farmers to academics. Amanzi for Food webpage
Kristen DiCerbo is breaking new ground in how games in the classroom can accelerate learning of all kinds. As Vice President of Learning Research and Design at Pearson, she’s on the forefront of defining and predicting what skills we will need for the future, such as how we can develop soft skills like creativity, collaboration, and critical thinking. For further information about the podcast and all the related links, visit www.aheadofourtime.com/learn-like-its-2030/.
This episode features an interview with Rigel Smiroldo recorded at NIPS 2017 in Long Beach California. We discuss data privacy, machine learning use cases, model deployment, and end-to-end machine learning.
Bryan Catanzaro, vice president for applied deep learning research at NVIDIA, talks about how we know an AI technology is working, the potential for AI-powered speech, and where we’ll see the next deep learning breakthroughs.
Mitchel Resnick (or Mitch, for short) knows his making—from a lot of different angles. And he’s not too bought into the whole “electronics and gadgets” side of the maker movement. Resnick has been in this business for more than 30 years, and it’s safe to say that he’s seen the maker movement—and the state of STEM education, in general—go through its phases, its ups and downs. He’s currently the LEGO Papert Professor of Learning Research and head of the Lifelong Kindergarten group at the MIT Media Lab, where he and his team have developed products familiar to many a science educator: the "programmable brick" technology that inspired the LEGO Mindstorms robotics kit, and Scratch, an online computing environment for students to learn about computer science. Is making something that every school should be doing—and are all interpretations of “making” of equitable value? EdSurge sat down with Resnick in his office at the MIT Media Lab to learn more, and to find out how he and his team are working to bring more creativity into the learning process.
GAMES IN EDUCATION PEARSON RESEARCH 's Dr Kristen DiCerbo who leads Pearson's who leads their Center for Learning Research and Technology is our guest
Connecting Advances in Learning Research and Teacher Practice: A Conference about Teacher Education Welcome and Conference Overview by Susan Fuhrman, President, Teachers College, Columbia University Conference Framing by Deborah Loewenberg Ball, William H. Payne Collegiate Chair; Arthur F. Thurnau Professor; Dean of the School of Education, University of Michigan
Erin Reilly is Research Director for Project New Media Literacies, a past CMS project now housed at the University of Southern California. Karen Schrier, a CMS grad, is the Director of Interactive Media and Technology at ESI Design and a part-time doctoral student at Columbia University in games and learning. Sangita Shresthova is a Czech/Nepali international development specialist, filmmaker, media scholar, and dancer, who currently manages Henry Jenkins new project on participatory culture and civic engagement at USC. Pilar Lacasa is a researcher at Alcalá University in Spain. She also works on a project for Electronic Arts in Spain about how to use commercial games in education. Mitch Resnick is Professor of Learning Research at the MIT Media Laboratory. He develops new technologies that engage children in creative learning experiences and is a principal investigator with the MIT Center for Future Civic Media, a CMS-partnered project.
Mitch Resnick, Professor of Learning Research at the MIT Media Lab and head of the Scratch development team introduces us to Scratch, a programming language that makes it easy to create interactive stories, animations, games, music, and art and share them on the Web.Scratch is designed to enhance the technological fluency of young people, helping them learn to express themselves creatively with new technologies.