Podcasts about tabnine

  • 34PODCASTS
  • 39EPISODES
  • 50mAVG DURATION
  • 1MONTHLY NEW EPISODE
  • Apr 29, 2025LATEST

POPULARITY

20172018201920202021202220232024


Best podcasts about tabnine

Latest podcast episodes about tabnine

Software Engineering Radio - The Podcast for Professional Software Developers
SE Radio 666: Eran Yahav on the Tabnine AI Coding Assistant

Software Engineering Radio - The Podcast for Professional Software Developers

Play Episode Listen Later Apr 29, 2025 62:05


Eran Yahav, Professor of Computer Science at Technion, Israel, and CTO of Tabnine, speaks with host Gregory M. Kapfhammer about the Tabnine AI coding assistant. They discuss how the design and implementation allows software engineers to use code completion and perform tasks such as automated code review while still maintaining developer privacy. Eran and Gregory also explore how research in the field of natural language processing (NLP) and large language models (LLMs) has informed the features in Tabnine. Brought to you by IEEE Computer Society and IEEE Software magazine.

Mingis on Tech
Will generative AI replace software engineers? | Ep. 222

Mingis on Tech

Play Episode Listen Later Apr 3, 2025 35:10


Generative AI has entered the world of software development—and it's making waves. In this episode of Today in Tech, host Keith Shaw is joined by Murali Sastry from Skillsoft and Eran Yanav from Tabnine to tackle the big question: Will generative AI replace mid-level software engineers? Or are developers evolving into AI-augmented leaders? From "vibe coding" and AI-generated pull request reviews to the future of entry-level coding jobs and computer science education, we explore how businesses are adapting and what skills developers need in this new era. :small_blue_diamond: Are companies still hiring coders? :small_blue_diamond: Is GenAI reliable for mission-critical code? :small_blue_diamond: What is “vibe coding,” and should you be worried? :small_blue_diamond: How is education shifting for the next generation of engineers? :point_right: Don't miss this deep dive into how AI is transforming the coding landscape. #GenerativeAI #SoftwareEngineering #TodayInTech #CodingWithAI #VibeCoding #AItools #TechTalk

What the Dev?
297: Why a clean codebase is key when using AI-assisted coding tools (with Tabnine's Eran Yahav)

What the Dev?

Play Episode Listen Later Feb 25, 2025 12:31


In this episode David Rubinstein interviews Eran Yahav, co-founder and CTO of Tabnine, about why its important to provide AI assistants with clean code.They discuss: The importance of defining organizational rules and best practices to guide the AIHow to use AI-assisted refactoring to improve legacy codebasesIf organizations should apply KonMari principles to their codebases

Enterprise Java Newscast
Stackd 75: He's a mystery man

Enterprise Java Newscast

Play Episode Listen Later Dec 4, 2024


Overview Josh, Kito, and Danno are joined by special guest Nilanjan Raychaudhuri, founder of Tublian and author of Scala in Action.They discuss the retirement of James Gosling, Jakarta EE 12, software development agents, TabNine, LangGraph for Java,...

Enterprise Java Newscast
Stackd 75: He's a mystery man

Enterprise Java Newscast

Play Episode Listen Later Dec 4, 2024 96:27


Overview Josh, Kito, and Danno are joined by special guest Nilanjan Raychaudhuri, founder of Tublian and author of Scala in Action.They discuss the retirement of James Gosling, Jakarta EE 12, software development agents, TabNine, LangGraph for Java, TinyLLM, NVIDIA's nvidia/Llama-3.1-Nemotron-70B-Instruct model, JDK 23, and Tublian's use of AI to empower the next generation of software developers. News  Industry News Java creator James Gosling officially announced his retirement and started a new chapter - ZOCNET White House Paper: Back to the Building Blocks: Path Toward Secure and Measurable Software Server Side Java Jakarta EE 12 Release Plan Draft AI/ML Tabnine AI agents generate, validate code for Jira issues | InfoWorld GitHub - bsorrentino/langgraph4j:

HTML All The Things - Web Development, Web Design, Small Business
Does AI Have Tech Bias? | AI All The Things

HTML All The Things - Web Development, Web Design, Small Business

Play Episode Listen Later Nov 26, 2024 61:41


In this episode, Matt and Mike introduce a new episode type dedicated to exploring the rapidly evolving world of AI. As AI tools advance at lightning speed, staying informed is critical for developers navigating this transformative era. This week, they dive into the evolution of AI-powered development tools, from simple autocompletes like Copilot and TabNine to full-scale IDE solutions like Cursor and Supermaven. They discuss the pros and cons of using these advanced tools for multifile code generation and manipulation, highlighting both the efficiencies and risks they bring. The conversation then shifts to the rise of AI-driven full-stack application generators, such as Bolt.new, V0, and GitHub Spark. These tools can build entire applications from simple prompts but come with a notable downside: tech bias. Matt shares his experience building a podcast website and highlights how AI's reliance on popular frameworks can limit the adoption of emerging technologies. Finally, the duo debates the future of AI in the development industry. Will AI replace developers in 1, 3, or even 10 years? Tune in to find out! Show Notes: https://www.htmlallthethings.com/podcasts/does-ai-have-tech-bias-ai-all-the-things Thanks to Wix Studio for sponsoring this episode! Check out Wix Studio, the web platform tailored to designers, developers, and marketers via this link: https://www.wix.com/studio

Dev Interrupted
How Specialized Models Drive Developer Productivity | Tabnine's Brandon Jung

Dev Interrupted

Play Episode Listen Later Sep 24, 2024 45:50 Transcription Available


What are the limitations of general large language models, and when should you evaluate more specialized models for your team's most important use case?This week, Conor Bronsdon sits down with Brandon Jung, Vice President of Ecosystem at Tabnine, to explore the difference between specialized models and LLMs. Brandon highlights how specialized models outperform LLMs when it comes to specific coding tasks, and how developers can leverage tailored solutions to improve developer productivity and code quality. The conversation covers the importance of data transparency, data origination, cost implications, and regulatory considerations such as the EU's AI Act.Whether you're a developer looking to boost your productivity or an engineering leader evaluating solutions for your team, this episode offers important context on the next wave of AI solutionsTopics:00:31 Specialized models vs. LLMs01:56 The problems with LLMs and data integrity12:34 Why AGI is further away than we think16:11 Evaluating the right models for your engineering team23:42 Is AI code secure?26:22 How to adjust to work with AI effectively 32:48 Training developers in the new AI worldLinks:Brandon Jung on LinkedInBrandon Jung (@brandoncjung) / XTabnine (@tabnine) / XTabnine AI code assistant | Private, personalized, protectedManaging Bot-Generated PRs & Reducing Team Workload by 6%Support the show: Subscribe to our Substack Leave us a review Subscribe on YouTube Follow us on Twitter or LinkedIn Offers: Learn about Continuous Merge with gitStream Get your DORA Metrics free forever

Eye On A.I.
#204 Peter Guagenti: How AI Will Change the Way Developers Work (Tabnine's Vision Explained)

Eye On A.I.

Play Episode Listen Later Aug 21, 2024 55:30


In this episode of the Eye on AI podcast, we sit down with Peter Guagenti, President and Chief Marketing Officer at Tabnine, to explore the role of AI in software development.   Peter takes us through his journey from web developer and engineering lead to leading Tabnine, a pioneer of AI code assistance.   We delve into the innovative ways Tabnine is pushing the boundaries of AI, from enhancing code generation to ensuring privacy with its Protected Model—offering enterprises fully private AI solutions tailored to their specific needs. Peter discusses how Tabnine is addressing the challenges of fit-to-purpose AI, making AI tools more context-aware and personalized to the workflows of individual engineering teams.   Peter also sheds light on the future of AI in software development, addressing the pressing question: Can AI truly replace developers, or is it destined to be a powerful collaborator?   Learn how AI can elevate software engineering teams, helping them overcome the repetitive tasks that slow down progress and focus on the creative aspects that push the industry forward.   Don't forget to like, subscribe, and hit the notification bell for more in-depth conversations on the latest AI advancements.     This episode of Eye on AI  is sponsored by BetterHelp. If you're thinking of starting therapy, give BetterHelp a try. It's entirely online. Designed to be convenient, flexible, and suited to your schedule. Just fill out a brief questionnaire to get matched with a licensed therapist, and switch therapists any time for no additional charge. Visit https://www.betterhelp.com/eyeonai today to get 10% off your first month.     Stay Updated: Craig Smith Twitter: https://twitter.com/craigss Eye on A.I. Twitter: https://twitter.com/EyeOn_AI     (00:00) Preview and Introduction (00:38) Peter Guagenti's Background (01:20) Tabnine's Origins (03:49) Innovating in AI Code Assistance (05:10) The Path to Autonomous Code Generation (07:49) Human Oversight in Autonomous AI (10:17) Misconceptions About AI Replacing Engineers (14:15) Future of Software Development with AI (17:04) Autonomous JIRA Tool and Broader Applications (22:36) Leveraging Vector Databases for Context (27:34) Balancing Contextual Data for AI (29:54) Expanding Generative AI Use Cases (34:16) Ensuring Code Quality with AI (37:17) Curating Quality Data for AI Models (41:17) The Need for Skilled Coders in AI (42:59) Future of Generative AI Beyond LLMs (47:00) Case Studies: Tabnine's Impact on Productivity (51:49) Conclusion: Building Trust in AI Technology

What the Dev?
266: Privacy concerns with Apple Intelligence and Apple's partnership with OpenAI (with Tabnine's Brandon Jung)

What the Dev?

Play Episode Listen Later Jun 25, 2024 16:05


In this episode, SD Times editor-in-chief David Rubinstein interviews Brandon Jung, VP of ecosystem and business development for Tabnine about his concerns with Apple's partnership with OpenAI, which was announced alongside Apple Intelligence as a way to pull in real-world knowledge from ChatGPT for questions Apple's own model cannot answer. They discuss: Why it's surprising that Apple decided to partner with OpenAIThe lack of transparency into Apple's own modelWhy open source models are better for innovation

HTML All The Things - Web Development, Web Design, Small Business
Will AI Replace Us? w/ The Creator of TabNine and Supermaven Jacob Jackson

HTML All The Things - Web Development, Web Design, Small Business

Play Episode Listen Later May 23, 2024 42:38


This week we had the pleasure of sitting down with Jacob Jackson, the creator of TabNine and Supermaven to discuss AI as a whole. It's no surprise that LLMs are taking over most of the current chatter in the tech world, but their rapid rise in popularity has also led to a rapid rise in concern. Many people believe that AI is coming for our jobs, working to replace human developers and other workers. On the other side of the fence, people are using AI to boost their productivity both at work and at home. Developers seem to be getting a great deal of this productivity boost with the creation of tools like Supermaven that can help speed up coding. If you've ever had any questions or doubt surrounding AI, or LLMs in general, then you're not going to want to miss this episode. Show Notes: https://www.htmlallthethings.com/podcasts/will-ai-replace-us-w-the-creator-of-tabnine-and-supermaven-jacob-jackson Learn with Scrimba: https://scrimba.com/?ref=htmlallthethings  

ai creator developers tabnine scrimba
Data Engineering Podcast
Build Your Second Brain One Piece At A Time

Data Engineering Podcast

Play Episode Listen Later Apr 28, 2024 50:10


Summary Generative AI promises to accelerate the productivity of human collaborators. Currently the primary way of working with these tools is through a conversational prompt, which is often cumbersome and unwieldy. In order to simplify the integration of AI capabilities into developer workflows Tsavo Knott helped create Pieces, a powerful collection of tools that complements the tools that developers already use. In this episode he explains the data collection and preparation process, the collection of model types and sizes that work together to power the experience, and how to incorporate it into your workflow to act as a second brain. Announcements Hello and welcome to the Data Engineering Podcast, the show about modern data management Dagster offers a new approach to building and running data platforms and data pipelines. It is an open-source, cloud-native orchestrator for the whole development lifecycle, with integrated lineage and observability, a declarative programming model, and best-in-class testability. Your team can get up and running in minutes thanks to Dagster Cloud, an enterprise-class hosted solution that offers serverless and hybrid deployments, enhanced security, and on-demand ephemeral test deployments. Go to dataengineeringpodcast.com/dagster (https://www.dataengineeringpodcast.com/dagster) today to get started. Your first 30 days are free! Data lakes are notoriously complex. For data engineers who battle to build and scale high quality data workflows on the data lake, Starburst powers petabyte-scale SQL analytics fast, at a fraction of the cost of traditional methods, so that you can meet all your data needs ranging from AI to data applications to complete analytics. Trusted by teams of all sizes, including Comcast and Doordash, Starburst is a data lake analytics platform that delivers the adaptability and flexibility a lakehouse ecosystem promises. And Starburst does all of this on an open architecture with first-class support for Apache Iceberg, Delta Lake and Hudi, so you always maintain ownership of your data. Want to see Starburst in action? Go to dataengineeringpodcast.com/starburst (https://www.dataengineeringpodcast.com/starburst) and get $500 in credits to try Starburst Galaxy today, the easiest and fastest way to get started using Trino. Your host is Tobias Macey and today I'm interviewing Tsavo Knott about Pieces, a personal AI toolkit to improve the efficiency of developers Interview Introduction How did you get involved in machine learning? Can you describe what Pieces is and the story behind it? The past few months have seen an endless series of personalized AI tools launched. What are the features and focus of Pieces that might encourage someone to use it over the alternatives? model selections architecture of Pieces application local vs. hybrid vs. online models model update/delivery process data preparation/serving for models in context of Pieces app application of AI to developer workflows types of workflows that people are building with pieces What are the most interesting, innovative, or unexpected ways that you have seen Pieces used? What are the most interesting, unexpected, or challenging lessons that you have learned while working on Pieces? When is Pieces the wrong choice? What do you have planned for the future of Pieces? Contact Info LinkedIn (https://www.linkedin.com/in/tsavoknott/) Parting Question From your perspective, what is the biggest barrier to adoption of machine learning today? Closing Announcements Thank you for listening! Don't forget to check out our other shows. Podcast.__init__ (https://www.pythonpodcast.com) covers the Python language, its community, and the innovative ways it is being used. The Machine Learning Podcast (https://www.themachinelearningpodcast.com) helps you go from idea to production with machine learning. Visit the site (https://www.dataengineeringpodcast.com) to subscribe to the show, sign up for the mailing list, and read the show notes. If you've learned something or tried out a project from the show then tell us about it! Email hosts@dataengineeringpodcast.com (mailto:hosts@dataengineeringpodcast.com)) with your story. Links Pieces (https://pieces.app/) NPU == Neural Processing Unit (https://en.wikipedia.org/wiki/AI_accelerator) Tensor Chip (https://en.wikipedia.org/wiki/Google_Tensor) LoRA == Low Rank Adaptation (https://github.com/microsoft/LoRA) Generative Adversarial Networks (https://en.wikipedia.org/wiki/Generative_adversarial_network) Mistral (https://mistral.ai/) Emacs (https://www.gnu.org/software/emacs/) Vim (https://www.vim.org/) NeoVim (https://neovim.io/) Dart (https://dart.dev/) Flutter (https://flutter.dev/) Typescript (https://www.typescriptlang.org/) Lua (https://www.lua.org/) Retrieval Augmented Generation (https://github.blog/2024-04-04-what-is-retrieval-augmented-generation-and-what-does-it-do-for-generative-ai/) ONNX (https://onnx.ai/) LSTM == Long Short-Term Memory (https://en.wikipedia.org/wiki/Long_short-term_memory) LLama 2 (https://llama.meta.com/llama2/) GitHub Copilot (https://github.com/features/copilot) Tabnine (https://www.tabnine.com/) Podcast Episode (https://www.themachinelearningpodcast.com/tabnine-generative-ai-developer-assistant-episode-24) The intro and outro music is from Hitman's Lovesong feat. Paola Graziano (https://freemusicarchive.org/music/The_Freak_Fandango_Orchestra/Tales_Of_A_Dead_Fish/Hitmans_Lovesong/) by The Freak Fandango Orchestra (http://freemusicarchive.org/music/The_Freak_Fandango_Orchestra/)/CC BY-SA 3.0 (https://creativecommons.org/licenses/by-sa/3.0/)

MLOps.community
What Business Stakeholders Want to See from the ML Teams // Peter Guagenti // #222

MLOps.community

Play Episode Listen Later Apr 2, 2024 81:27


Join us at our first in-person conference on June 25 all about AI Quality: https://www.aiqualityconference.com/ ⁠Peter Guagenti⁠ is an accomplished business builder and entrepreneur with expertise in strategy, product development, marketing, sales, and operations. Peter has helped build multiple successful start-ups to exits, fueling high growth in each company along the way. He brings a broad perspective, deep problem-solving skills, the ability to drive innovation amongst teams, and a proven ability to convert strategy into action -- all backed up by a history of delivering results. Huge thank you to AWS for sponsoring this episode. AWS - https://aws.amazon.com/ MLOps podcast #222 with Peter Guagenti, President & CMO of Tabnine - What Business Stakeholders Want to See from the ML Teams. // Abstract Peter Guagenti shares his expertise in the tech industry, discussing topics from managing large-scale tech legacy applications and data experimentation to the evolution of the Internet. He returns to his history of building and transforming businesses, such as his work in the early 90s for People magazine's website and his current involvement in AI development for software companies. Guagenti discusses the use of predictive modeling in customer management and emphasizes the importance of re-architecting solutions to fit customer needs. He also delves deeper into the AI tools' effectiveness in software development and the value of maintaining privacy. Guagenti sees a bright future in AI democratization and shares his company's development of AI coding assistants. Discussing successful entrepreneurship, Guagenti highlights balancing technology and go-to-market strategies and the value of failing fast. // Bio Peter Guagenti is the President and Chief Marketing Officer at Tabnine. Guagenti is an accomplished business leader and entrepreneur with expertise in strategy, product development, marketing, sales, and operations. He most recently served as chief marketing officer at Cockroach Labs, and he previously held leadership positions at SingleStore, NGINX (acquired by F5 Networks), and Acquia (acquired by Vista Equity Partners). Guagenti also serves as an advisor to a number of visionary AI and data companies including DragonflyDB, Memgraph, and Treeverse. // MLOps Jobs board https://mlops.pallet.xyz/jobs // MLOps Swag/Merch https://mlops-community.myshopify.com/ // Related Links AI Quality in Person Conference: https://www.aiqualityconference.com/ Measuring the impact of GitHub Copilot Survey: https://resources.github.com/learn/pathways/copilot/essentials/measuring-the-impact-of-github-copilot/ AWS Trainium and Inferentia: https://aws.amazon.com/machine-learning/trainium/ https://aws.amazon.com/machine-learning/inferentia/AI coding assistants: 8 features enterprises should seek: https://www.infoworld.com/article/3694900/ai-coding-assistants-8-features-enterprises-should-seek.htmlCareers at Tabnine: https://www.tabnine.com/careers --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Catch all episodes, blogs, newsletters, and more: https://mlops.community/ Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Peter on LinkedIn: https://www.linkedin.com/in/peterguagenti/

10KMedia Podcast
Episode 42: Peter Guagenti, President & CMO at Tabnine

10KMedia Podcast

Play Episode Listen Later Feb 6, 2024 46:12


Adam sits down with Peter to discuss how coding assistants can democratize knowledge and boost productivity, the potential copyright concerns with Generative AI, and if ultimately humans have much to worry about with the rise of AI.

ai tabnine
0800-DEVOPS
AI code assistants with Peter Guagenti

0800-DEVOPS

Play Episode Listen Later Feb 3, 2024 42:00


Peter Guagenti is the President & Chief Marketing Officer of Tabnine, a company behind one of the most popular AI code assistants on the market. I spoke with Peter about AI code assistants, what automatic transmissions and AI have in common, and the philosophical question of future humans forgetting how to code.Please leave a review on your favorite podcast platform or Podchaser, and subscribe to 0800-DEVOPS newsletter here.This interview is featured in 0800-DEVOPS #58 - AI code assistants with Peter Guagenti.[Check out podcast chapters if available on your podcast platform or use links below](0:45)Introduction(10:41)The magic behind AI code assistants(14:03)Typical patterns how developers use code assistants(21:15)Can developers rely too much on code assistants?(30:26)Will developer skills deteriorate over time?(34:52)How mature is the technology today?(39:22)The future of code assistants

Odbita do bita
Matej, Maruša in Anže o temah, ki so jih izbrali poslušalci

Odbita do bita

Play Episode Listen Later Nov 16, 2023 34:14


Slike, videe in besedila že prepričljivo generira umetna inteligenca. Kako in komu še zaupati? Katera orodja umetne inteligence vendarle prinašajo lažji študij ali delo v pisarni? Natanko eno leto po prvih omembah ChatGPT-ja se predvsem sprašujemo o možnostih uporabe in zlorabe. So ruske tehnologije nevarne? So telefoni na preklop modni trend ali korak v novo smer?Maruša in Anže debatirata s kolegom Matejem Praprotnikom o temah, ki so jih izbrali odbiti poslušalci. Zapiski: DeepL Translate: The world’s most accurate translator VERAS | EvolveLAB Stable Diffusion Online - AI Image Generator GitHub Copilot · Your AI pair programmer · GitHub Tabnine is an AI assistant that speeds up delivery and keeps your code safe CodeGeeX GitHub Copilot · Your AI pair programmer · GitHub Codeium · Free AI Code Completion & Chat Perplexity AI for Research - scite.ai ChatPDF - Chat with any PDF! Consensus: AI Search Engine for Research Chat LlamaIndex HuggingChat Razpravi o odbitih temah se lahko pridružite na Discordu.  

Data Engineering Podcast
Enhancing The Abilities Of Software Engineers With Generative AI At Tabnine

Data Engineering Podcast

Play Episode Listen Later Nov 13, 2023 67:52


Summary Software development involves an interesting balance of creativity and repetition of patterns. Generative AI has accelerated the ability of developer tools to provide useful suggestions that speed up the work of engineers. Tabnine is one of the main platforms offering an AI powered assistant for software engineers. In this episode Eran Yahav shares the journey that he has taken in building this product and the ways that it enhances the ability of humans to get their work done, and when the humans have to adapt to the tool. Announcements Hello and welcome to the Data Engineering Podcast, the show about modern data management Introducing RudderStack Profiles. RudderStack Profiles takes the SaaS guesswork and SQL grunt work out of building complete customer profiles so you can quickly ship actionable, enriched data to every downstream team. You specify the customer traits, then Profiles runs the joins and computations for you to create complete customer profiles. Get all of the details and try the new product today at dataengineeringpodcast.com/rudderstack (https://www.dataengineeringpodcast.com/rudderstack) This episode is brought to you by Datafold – a testing automation platform for data engineers that finds data quality issues before the code and data are deployed to production. Datafold leverages data-diffing to compare production and development environments and column-level lineage to show you the exact impact of every code change on data, metrics, and BI tools, keeping your team productive and stakeholders happy. Datafold integrates with dbt, the modern data stack, and seamlessly plugs in your data CI for team-wide and automated testing. If you are migrating to a modern data stack, Datafold can also help you automate data and code validation to speed up the migration. Learn more about Datafold by visiting dataengineeringpodcast.com/datafold (https://www.dataengineeringpodcast.com/datafold) Data lakes are notoriously complex. For data engineers who battle to build and scale high quality data workflows on the data lake, Starburst powers petabyte-scale SQL analytics fast, at a fraction of the cost of traditional methods, so that you can meet all your data needs ranging from AI to data applications to complete analytics. Trusted by teams of all sizes, including Comcast and Doordash, Starburst is a data lake analytics platform that delivers the adaptability and flexibility a lakehouse ecosystem promises. And Starburst does all of this on an open architecture with first-class support for Apache Iceberg, Delta Lake and Hudi, so you always maintain ownership of your data. Want to see Starburst in action? Go to dataengineeringpodcast.com/starburst (https://www.dataengineeringpodcast.com/starburst) and get $500 in credits to try Starburst Galaxy today, the easiest and fastest way to get started using Trino. You shouldn't have to throw away the database to build with fast-changing data. You should be able to keep the familiarity of SQL and the proven architecture of cloud warehouses, but swap the decades-old batch computation model for an efficient incremental engine to get complex queries that are always up-to-date. With Materialize, you can! It's the only true SQL streaming database built from the ground up to meet the needs of modern data products. Whether it's real-time dashboarding and analytics, personalization and segmentation or automation and alerting, Materialize gives you the ability to work with fresh, correct, and scalable results — all in a familiar SQL interface. Go to dataengineeringpodcast.com/materialize (https://www.dataengineeringpodcast.com/materialize) today to get 2 weeks free! Your host is Tobias Macey and today I'm interviewing Eran Yahav about building an AI powered developer assistant at Tabnine Interview Introduction How did you get involved in machine learning? Can you describe what Tabnine is and the story behind it? What are the individual and organizational motivations for using AI to generate code? What are the real-world limitations of generative AI for creating software? (e.g. size/complexity of the outputs, naming conventions, etc.) What are the elements of skepticism/oversight that developers need to exercise while using a system like Tabnine? What are some of the primary ways that developers interact with Tabnine during their development workflow? Are there any particular styles of software for which an AI is more appropriate/capable? (e.g. webapps vs. data pipelines vs. exploratory analysis, etc.) For natural languages there is a strong bias toward English in the current generation of LLMs. How does that translate into computer languages? (e.g. Python, Java, C++, etc.) Can you describe the structure and implementation of Tabnine? Do you rely primarily on a single core model, or do you have multiple models with subspecialization? How have the design and goals of the product changed since you first started working on it? What are the biggest challenges in building a custom LLM for code? What are the opportunities for specialization of the model architecture given the highly structured nature of the problem domain? For users of Tabnine, how do you assess/monitor the accuracy of recommendations? What are the feedback and reinforcement mechanisms for the model(s)? What are the most interesting, innovative, or unexpected ways that you have seen Tabnine's LLM powered coding assistant used? What are the most interesting, unexpected, or challenging lessons that you have learned while working on AI assisted development at Tabnine? When is an AI developer assistant the wrong choice? What do you have planned for the future of Tabnine? Contact Info LinkedIn (https://www.linkedin.com/in/eranyahav/?originalSubdomain=il) Website (https://csaws.cs.technion.ac.il/~yahave/) Parting Question From your perspective, what is the biggest barrier to adoption of machine learning today? Closing Announcements Thank you for listening! Don't forget to check out our other shows. Podcast.__init__ (https://www.pythonpodcast.com) covers the Python language, its community, and the innovative ways it is being used. The Machine Learning Podcast (https://www.themachinelearningpodcast.com) helps you go from idea to production with machine learning. Visit the site (https://www.dataengineeringpodcast.com) to subscribe to the show, sign up for the mailing list, and read the show notes. If you've learned something or tried out a project from the show then tell us about it! Email hosts@dataengineeringpodcast.com (mailto:hosts@dataengineeringpodcast.com)) with your story. To help other people find the show please leave a review on Apple Podcasts (https://podcasts.apple.com/us/podcast/data-engineering-podcast/id1193040557) and tell your friends and co-workers Links TabNine (https://www.tabnine.com/) Technion University (https://www.technion.ac.il/en/home-2/) Program Synthesis (https://en.wikipedia.org/wiki/Program_synthesis) Context Stuffing (http://gptprompts.wikidot.com/context-stuffing) Elixir (https://elixir-lang.org/) Dependency Injection (https://en.wikipedia.org/wiki/Dependency_injection) COBOL (https://en.wikipedia.org/wiki/COBOL) Verilog (https://en.wikipedia.org/wiki/Verilog) MidJourney (https://www.midjourney.com/home) The intro and outro music is from Hitman's Lovesong feat. Paola Graziano (https://freemusicarchive.org/music/The_Freak_Fandango_Orchestra/Tales_Of_A_Dead_Fish/Hitmans_Lovesong/) by The Freak Fandango Orchestra (http://freemusicarchive.org/music/The_Freak_Fandango_Orchestra/)/CC BY-SA 3.0 (https://creativecommons.org/licenses/by-sa/3.0/)

The Machine Learning Podcast
Enhancing The Abilities Of Software Engineers With Generative AI At Tabnine

The Machine Learning Podcast

Play Episode Listen Later Nov 13, 2023 64:47


Summary Software development involves an interesting balance of creativity and repetition of patterns. Generative AI has accelerated the ability of developer tools to provide useful suggestions that speed up the work of engineers. Tabnine is one of the main platforms offering an AI powered assistant for software engineers. In this episode Eran Yahav shares the journey that he has taken in building this product and the ways that it enhances the ability of humans to get their work done, and when the humans have to adapt to the tool. Announcements Hello and welcome to the Machine Learning Podcast, the podcast about machine learning and how to bring it from idea to delivery. Your host is Tobias Macey and today I'm interviewing Eran Yahav about building an AI powered developer assistant at Tabnine Interview Introduction How did you get involved in machine learning? Can you describe what Tabnine is and the story behind it? What are the individual and organizational motivations for using AI to generate code? What are the real-world limitations of generative AI for creating software? (e.g. size/complexity of the outputs, naming conventions, etc.) What are the elements of skepticism/oversight that developers need to exercise while using a system like Tabnine? What are some of the primary ways that developers interact with Tabnine during their development workflow? Are there any particular styles of software for which an AI is more appropriate/capable? (e.g. webapps vs. data pipelines vs. exploratory analysis, etc.) For natural languages there is a strong bias toward English in the current generation of LLMs. How does that translate into computer languages? (e.g. Python, Java, C++, etc.) Can you describe the structure and implementation of Tabnine? Do you rely primarily on a single core model, or do you have multiple models with subspecialization? How have the design and goals of the product changed since you first started working on it? What are the biggest challenges in building a custom LLM for code? What are the opportunities for specialization of the model architecture given the highly structured nature of the problem domain? For users of Tabnine, how do you assess/monitor the accuracy of recommendations? What are the feedback and reinforcement mechanisms for the model(s)? What are the most interesting, innovative, or unexpected ways that you have seen Tabnine's LLM powered coding assistant used? What are the most interesting, unexpected, or challenging lessons that you have learned while working on AI assisted development at Tabnine? When is an AI developer assistant the wrong choice? What do you have planned for the future of Tabnine? Contact Info LinkedIn (https://www.linkedin.com/in/eranyahav/?originalSubdomain=il) Website (https://csaws.cs.technion.ac.il/~yahave/) Parting Question From your perspective, what is the biggest barrier to adoption of machine learning today? Closing Announcements Thank you for listening! Don't forget to check out our other shows. The Data Engineering Podcast (https://www.dataengineeringpodcast.com) covers the latest on modern data management. Podcast.__init__ () covers the Python language, its community, and the innovative ways it is being used. Visit the site (https://www.themachinelearningpodcast.com) to subscribe to the show, sign up for the mailing list, and read the show notes. If you've learned something or tried out a project from the show then tell us about it! Email hosts@themachinelearningpodcast.com (mailto:hosts@themachinelearningpodcast.com)) with your story. To help other people find the show please leave a review on iTunes (https://podcasts.apple.com/us/podcast/the-machine-learning-podcast/id1626358243) and tell your friends and co-workers. Links TabNine (https://www.tabnine.com/) Technion University (https://www.technion.ac.il/en/home-2/) Program Synthesis (https://en.wikipedia.org/wiki/Program_synthesis) Context Stuffing (http://gptprompts.wikidot.com/context-stuffing) Elixir (https://elixir-lang.org/) Dependency Injection (https://en.wikipedia.org/wiki/Dependency_injection) COBOL (https://en.wikipedia.org/wiki/COBOL) Verilog (https://en.wikipedia.org/wiki/Verilog) MidJourney (https://www.midjourney.com/home) The intro and outro music is from Hitman's Lovesong feat. Paola Graziano (https://freemusicarchive.org/music/The_Freak_Fandango_Orchestra/Tales_Of_A_Dead_Fish/Hitmans_Lovesong/) by The Freak Fandango Orchestra (http://freemusicarchive.org/music/The_Freak_Fandango_Orchestra/)/CC BY-SA 3.0 (https://creativecommons.org/licenses/by-sa/3.0/)

Infinite Machine Learning
AI assistant for software development | Eran Yahav, cofounder and CTO of Tabnine

Infinite Machine Learning

Play Episode Listen Later Aug 21, 2023 39:34


Eran Yahav is the cofounder and CTO of Tabnine, an AI assistant that developers can use to build software faster. He's a professor at Technion - Israel Institute of Technology and was previously a researcher at IBM. He has a PhD in Computer Science from Tel Aviv University. In this episode, we cover a range of topics including: - Tasks in software development - What tasks are likely to benefit from LLMs - The launch of Tabnine Chat - Characteristics of a good AI coding assistant - Making AI coding assistants context-aware - Generic LLMs vs domain specific LLMs - AI copilot for devops work Eran's favorite book: Catch-22 (Author: Joseph Heller)--------Where to find Prateek Joshi: Newsletter: https://prateekjoshi.substack.com Website: https://prateekj.com LinkedIn: https://www.linkedin.com/in/prateek-joshi-91047b19 Twitter: https://twitter.com/prateekvjoshi 

Lenny's Podcast: Product | Growth | Career
How to measure and improve developer productivity | Nicole Forsgren (Microsoft Research, GitHub, Google)

Lenny's Podcast: Product | Growth | Career

Play Episode Listen Later Jul 30, 2023 76:17


This episode is brought to you by DX—a platform for measuring and improving developer productivity.—Dr. Nicole Forsgren is a developer productivity and DevOps expert who works with engineering organizations to make work better. Best known as co-author of the Shingo Publication Award-winning book Accelerate and the DevOps Handbook, 2nd edition and author of the State of DevOps Reports, she has helped some of the biggest companies in the world transform their culture, processes, tech, and architecture. Nicole is currently a Partner at Microsoft Research, leading developer productivity research and strategy, and a technical founder/CEO with a successful exit to Google. In a previous life, she was a software engineer, sysadmin, hardware performance engineer, and professor. She has published several peer-reviewed journal papers, has been awarded public and private research grants (funders include NASA and the NSF), and has been featured in the Wall Street Journal, Forbes, Computerworld, and InformationWeek. In today's podcast, we discuss:• Two frameworks for measuring developer productivity: DORA and SPACE• Benchmarks for what good and great look like• Common mistakes to avoid when measuring developer productivity• Resources and tools for improving your metrics• Signs your developer experience needs attention• How to improve your developer experience• Nicole's Four-Box framework for thinking about data and relationships—Find the full transcript at: https://www.lennyspodcast.com/how-to-measure-and-improve-developer-productivity-nicole-forsgren-microsoft-research-github-goo/#transcript—Where to find Nicole Forsgren:• Twitter: https://twitter.com/nicolefv• LinkedIn: https://www.linkedin.com/in/nicolefv/• Website: https://nicolefv.com/—Where to find Lenny:• Newsletter: https://www.lennysnewsletter.com• Twitter: https://twitter.com/lennysan• LinkedIn: https://www.linkedin.com/in/lennyrachitsky/—In this episode, we cover:(00:00) Nicole's background(07:55) Unpacking the terms “developer productivity,” “developer experience,” and “DevOps”(10:06) How to move faster and improve practices across the board(13:43) The DORA framework(18:54) Benchmarks for success(22:33) Why company size doesn't matter (24:54) How to improve DevOps capabilities by working backward(29:23) The SPACE framework and choosing metrics(32:51) How SPACE and DORA work together(35:39) Measuring satisfaction(37:52) Resources and tools for optimizing metrics(41:29) Nicole's current book project(45:43) Common pitfalls companies run into when rolling out developer productivity/optimizations(47:42) How the DevOps space has progressed(50:07) The impact of AI on the developer experience and productivity(54:04) First steps to take if you're trying to improve the developer experience(55:15) Why Google is an example of a company implementing DevOps solutions well(56:11) The importance of clear communication(57:32) Nicole's Four-Box framework(1:05:15) Advice on making decisions (1:08:56) Lightning round—Referenced:• Chef: https://www.chef.io/• DORA: https://dora.dev/• GitHub: https://github.com/• Microsoft Research: https://www.microsoft.com/en-us/research/• What is DORA?: https://devops.com/what-is-dora-and-why-you-should-care/• Dustin Smith on LinkedIn: https://www.linkedin.com/in/dustin-smith-b0525458/• Nathen Harvey on LinkedIn: https://www.linkedin.com/in/nathen/• What is CI/CD?: https://about.gitlab.com/topics/ci-cd/• Trunk-based development: https://cloud.google.com/architecture/devops/devops-tech-trunk-based-development• DORA DevOps Quick Check: https://dora.dev/quickcheck/• Accelerate: The Science of Lean Software and DevOps: Building and Scaling High Performing Technology Organizations: https://www.amazon.com/Accelerate-Software-Performing-Technology-Organizations/dp/1942788339• The SPACE of Developer Productivity: https://queue.acm.org/detail.cfm?id=3454124• DevOps Metrics: Nicole Forsgren and Mik Kersten: https://queue.acm.org/detail.cfm?id=3182626• How to Measure Anything: Finding the Value of Intangibles in Business: https://www.amazon.com/How-Measure-Anything-Intangibles-Business/dp/1118539273/• GitHub Copilot: https://github.com/features/copilot• Tabnine: https://www.tabnine.com/the-leading-ai-assistant-for-software-development• Nicole's Decision-Making Spreadsheet: https://docs.google.com/spreadsheets/d/1wItAODkhZ-zKnnFbyDERCd8Hq2NQ03WPvCfigBQ5vpc/edit?usp=sharing• How to do linear regression and correlation analysis: https://www.lennysnewsletter.com/p/linear-regression-and-correlation-analysis• Good Strategy/Bad Strategy: The difference and why it matters: https://www.amazon.com/Good-Strategy-Bad-difference-matters/dp/1781256179/• Designing Your Life: How to Build a Well-Lived, Joyful Life: https://www.amazon.com/Designing-Your-Life-Well-Lived-Joyful/dp/1101875321• Ender's Game: https://www.amazon.com/Enders-Game-Ender-Quintet-1/dp/1250773024/ref=tmm_pap_swatch_0• Suits on Netflix: https://www.netflix.com/title/70195800• Ted Lasso on AppleTV+: https://tv.apple.com/us/show/ted-lasso• Never Have I Ever on Netflix: https://www.netflix.com/title/80179190• Eight Sleep: https://www.eightsleep.com/• COSRX face masks: https://www.amazon.com/COSRX-Advanced-Secretion-Hydrating-Moisturizing/dp/B08JSL9W6K/—Production and marketing by https://penname.co/. For inquiries about sponsoring the podcast, email podcast@lennyrachitsky.com.—Lenny may be an investor in the companies discussed. Get full access to Lenny's Newsletter at www.lennysnewsletter.com/subscribe

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0
The AI Founder Gene: Being Early, Building Fast, and Believing in Greatness — with Sharif Shameem of Lexica

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

Play Episode Listen Later May 8, 2023 50:37


Thanks to the over 42,000 latent space explorers who checked out our Replit episode! We are hosting/attending a couple more events in SF and NYC this month. See you if in town!Lexica.art was introduced to the world 24 hours after the release of Stable Diffusion as a search engine for prompts, gaining instant product-market fit as a world discovering generative AI also found they needed to learn prompting by example.Lexica is now 8 months old, serving 5B image searches/day, and just shipped V3 of Lexica Aperture, their own text-to-image model! Sharif Shameem breaks his podcast hiatus with us for an exclusive interview covering his journey building everything with AI!The conversation is nominally about Sharif's journey through his three startups VectorDash, Debuild, and now Lexica, but really a deeper introspection into what it takes to be a top founder in the fastest moving tech startup scene (possibly ever) of AI. We hope you enjoy this conversation as much as we did!Full transcript is below the fold. We would really appreciate if you shared our pod with friends on Twitter, LinkedIn, Mastodon, Bluesky, or your social media poison of choice!Timestamps* [00:00] Introducing Sharif* [02:00] VectorDash* [05:00] The GPT3 Moment and Building Debuild* [09:00] Stable Diffusion and Lexica* [11:00] Lexica's Launch & How it Works* [15:00] Being Chronically Early* [16:00] From Search to Custom Models* [17:00] AI Grant Learnings* [19:30] The Text to Image Illuminati?* [20:30] How to Learn to Train Models* [24:00] The future of Agents and Human Intervention* [29:30] GPT4 and Multimodality* [33:30] Sharif's Startup Manual* [38:30] Lexica Aperture V1/2/3* [40:00] Request for AI Startup - LLM Tools* [41:00] Sequencing your Genome* [42:00] Believe in Doing Great Things* [44:30] Lightning RoundShow Notes* Sharif's website, Twitter, LinkedIn* VectorDash (5x cheaper than AWS)* Debuild Insider, Fast company, MIT review, tweet, tweet* Lexica* Introducing Lexica* Lexica Stats* Aug: “God mode” search* Sep: Lexica API * Sept: Search engine with CLIP * Sept: Reverse image search* Nov: teasing Aperture* Dec: Aperture v1* Dec - Aperture v2* Jan 2023 - Outpainting* Apr 2023 - Aperture v3* Same.energy* AI Grant* Sharif on Agents: prescient Airpods tweet, Reflection* MiniGPT4 - Sharif on Multimodality* Sharif Startup Manual* Sharif Future* 23andMe Genome Sequencing Tool: Promethease* Lightning Round* Fave AI Product: Cursor.so. Swyx ChatGPT Menubar App.* Acceleration: Multimodality of GPT4. Animated Drawings* Request for Startup: Tools for LLMs, Brex for GPT Agents* Message: Build Weird Ideas!TranscriptAlessio: Hey everyone, welcome to the Latent Space podcast. This is Alessio, partner and CTO on Residence at Decibel Partners. I'm joined by my co-host Wix, writer and editor of Latent Space. And today we have Sharish Amin. Welcome to the studio. Sharif: Awesome. Thanks for the invite.Swyx: Really glad to have you. [00:00] Introducing SharifSwyx: You've been a dream guest, actually, since we started drafting guest lists for this pod. So glad we could finally make this happen. So what I like to do is usually introduce people, offer their LinkedIn, and then prompt you for what's not on your LinkedIn. And to get a little bit of the person behind the awesome projects. So you graduated University of Maryland in CS. Sharif: So I actually didn't graduate, but I did study. Swyx: You did not graduate. You dropped out. Sharif: I did drop out. Swyx: What was the decision behind dropping out? Sharif: So first of all, I wasn't doing too well in any of my classes. I was working on a side project that took up most of my time. Then I spoke to this guy who ended up being one of our investors. And he was like, actually, I ended up dropping out. I did YC. And my company didn't end up working out. And I returned to school and graduated along with my friends. I was like, oh, it's actually a reversible decision. And that was like that. And then I read this book called The Case Against Education by Brian Kaplan. So those two things kind of sealed the deal for me on dropping out. Swyx: Are you still on hiatus? Could you still theoretically go back? Sharif: Theoretically, probably. Yeah. Still on indefinite leave. Swyx: Then you did some work at Mitra? Sharif: Mitra, yeah. So they're lesser known. So they're technically like an FFRDC, a federally funded research and development center. So they're kind of like a large government contractor, but nonprofit. Yeah, I did some computer vision work there as well. [02:00] VectorDashSwyx: But it seems like you always have an independent founder bone in you. Because then you started working on VectorDash, which is distributed GPUs. Sharif: Yes. Yeah. So VectorDash was a really fun project that we ended up working on for a while. So while I was at Mitra, I had a friend who was mining Ethereum. This was, I think, 2016 or 2017. Oh my God. Yeah. And he was mining on his NVIDIA 1080Ti, making around like five or six dollars a day. And I was trying to train a character recurrent neural network, like a character RNN on my iMessage text messages to make it like a chatbot. Because I was just curious if I could do it. Because iMessage stores all your past messages from years ago in a SQL database, which is pretty nifty. But I wanted to train it. And I needed a GPU. And it was, I think, $60 to $80 for a T4 on AWS, which is really slow compared to a 1080Ti. If you normalize the cost and performance versus the 1080Ti when someone's mining Ethereum, it's like a 20x difference. So I was like, hey, his name was Alex. Alex, I'll give you like 10 bucks if you let me borrow your 1080Ti for a week. I'll give you 10 bucks per day. And it was like 70 bucks. And I used it to train my model. And it worked great. The model was really bad, but the whole trade worked really great. I got a really high performance GPU to train my model on. He got much more than he was making by mining Ethereum. So we had this idea. I was like, hey, what if we built this marketplace where people could rent their GPUs where they're mining cryptocurrency and machine learning researchers could just rent them out and pay a lot cheaper than they would pay AWS. And it worked pretty well. We launched in a few months. We had over 120,000 NVIDIA GPUs on the platform. And then we were the cheapest GPU cloud provider for like a solid year or so. You could rent a pretty solid GPU for like 20 cents an hour. And cryptocurrency miners were making more than they would make mining crypto because this was after the Ethereum crash. And yeah, it was pretty cool. It just turns out that a lot of our customers were college students and researchers who didn't have much money. And they weren't necessarily the best customers to have as a business. Startups had a ton of credits and larger companies were like, actually, we don't really trust you with our data, which makes sense. Yeah, we ended up pivoting that to becoming a cloud GPU provider for video games. So we would stream games from our GPUs. Oftentimes, like many were located just a few blocks away from you because we had the lowest latency of any cloud GPU provider, even lower than like AWS and sometimes Cloudflare. And we decided to build a cloud gaming platform where you could pretty much play your own games on the GPU and then stream it back to your Mac or PC. Swyx: So Stadia before Stadia. Sharif: Yeah, Stadia before Stadia. It's like a year or so before Stadia. Swtx: Wow. Weren't you jealous of, I mean, I don't know, it sounds like Stadia could have bought you or Google could have bought you for Stadia and that never happened? Sharif: It never happened. Yeah, it didn't end up working out for a few reasons. The biggest thing was internet bandwidth. So a lot of the hosts, the GPU hosts had lots of GPUs, but average upload bandwidth in the United States is only 35 megabits per second, I think. And like a 4K stream needs like a minimum of 15 to 20 megabits per second. So you could really only utilize one of those GPUs, even if they had like 60 or 100. [05:00] The GPT3 Moment and Building DebuildSwyx: And then you went to debuild July 2020, is the date that I have. I'm actually kind of just curious, like what was your GPT-3 aha moment? When were you like GPT-3-pilled? Sharif: Okay, so I first heard about it because I was also working on another chatbot. So this was like after, like everything ties back to this chatbot I'm trying to make. This was after working on VectorDash. I was just like hacking on random projects. I wanted to make the chatbot using not really GPT-2, but rather just like it would be pre-programmed. It was pretty much you would give it a goal and then it would ask you throughout the week how much progress you're making to that goal. So take your unstructured response, usually a reply to a text message, and then it would like, plot it for you in like a table and you could see your progress over time. It could be for running or tracking calories. But I wanted to use GPT-3 to make it seem more natural because I remember someone on Bookface, which is still YC's internal forum. They posted and they were like, OpenAI just released AGI and it's GPT-3. I asked it like a bunch of logic puzzles and it solved them all perfectly. And I was like, what? How's no one else talking about this? Like this is either like the greatest thing ever that everyone is missing or like it's not that good. So like I tweeted out if anyone could get me access to it. A few hours later, Greg Brockman responded. Swyx: He is everywhere. Sharif: He's great. Yeah, he's on top of things. And yeah, by that afternoon, I was like messing around with the API and I was like, wow, this is incredible. You could chat with fake people or people that have passed away. You could like, I remember the first conversation I did was this is a chat with Steve Jobs and it was like, interviewer, hi. What are you up to today on Steve? And then like you could talk to Steve Jobs and it was somewhat plausible. Oh, the thing that really blew my mind was I tried to generate code with it. So I'd write the function for a JavaScript header or the header for a JavaScript function. And it would complete the rest of the function. I was like, whoa, does this code actually work? Like I copied it and ran it and it worked. And I tried it again. I gave more complex things and like I kind of understood where it would break, which was like if it was like something, like if it was something you couldn't easily describe in a sentence and like contain all the logic for in a single sentence. So I wanted to build a way where I could visually test whether these functions were actually working. And what I was doing was like I was generating the code in the playground, copying it into my VS code editor, running it and then reloading the react development page. And I was like, okay, cool. That works. So I was like, wait, let me just put this all in like the same page so I can just compile in the browser, run it in the browser and then submit it to the API in the browser as well. So I did that. And it was really just like a simple loop where you just type in the prompt. It would generate the code and then compile it directly in the browser. And it showed you the response. And I did this for like very basic JSX react components. I mean, it worked. It was pretty mind blowing. I remember staying up all night, like working on it. And it was like the coolest thing I'd ever worked on at the time so far. Yeah. And then I was like so mind blowing that no one was talking about this whole GPT three thing. I was like, why is this not on everyone's minds? So I recorded a quick 30 second demo and I posted on Twitter and like I go to bed after staying awake for like 20 hours straight. When I wake up the next morning and I had like 20,000 likes and like 100,000 people had viewed it. I was like, oh, this is so cool. And then I just kept putting demos out for like the next week. And yeah, that was like my GPT three spark moment. Swyx: And you got featured in like Fast Company, MIT Tech Review, you know, a bunch of stuff, right? Sharif: Yeah. Yeah. I think a lot of it was just like the API had been there for like a month prior already. Swyx: Not everyone had access. Sharif: That's true. Not everyone had access. Swyx: So you just had the gumption to tweet it out. And obviously, Greg, you know, on top of things as always. Sharif: Yeah. Yeah. I think it also makes a lot of sense when you kind of share things in a way that's easily consumable for people to understand. Whereas if you had shown a terminal screenshot of a generating code, that'd be pretty compelling. But whereas seeing it get rendered and compiled directly in front of you, there's a lot more interesting. There's also that human aspect to it where you want to relate things to the end user, not just like no one really cares about evals. When you can create a much more compelling demo explaining how it does on certain tasks. [09:00] Stable Diffusion and LexicaSwyx: Okay. We'll round it out soon. But in 2022, you moved from Debuild to Lexica, which was the search engine. I assume this was inspired by stable diffusion, but I can get the history there a little bit. Sharif: Yeah. So I was still working on Debuild. We were growing at like a modest pace and I was in the stable... Swyx: I was on the signup list. I never got off. Sharif: Oh yeah. Well, we'll get you off. It's not getting many updates anymore, but yeah, I was in the stable diffusion discord and I was in it for like many hours a day. It was just like the most exciting thing I'd ever done in a discord. It was so cool. Like people were generating so many images, but I didn't really know how to write prompts and people were like writing really complicated things. They would be like, like a modern home training on our station by Greg Rutkowski, like a 4k Unreal Engine. It's like that there's no way that actually makes the images look better. But everyone was just kind of copying everyone else's prompts and like changing like the first few words. Swyx: Yeah. Yeah. Sharif: So I was like using the discord search bar and it was really bad because it showed like five images at a time. And I was like, you know what? I could build a much better interface for this. So I ended up scraping the entire discord. It was like 10 million images. I put them in a database and I just pretty much built a very basic search engine where you could just type for type a word and then it returned all the prompts that had that word. And I built the entire website for it in like 20, in like about two days. And we shipped it the day I shipped it the day after the stable diffusion weights were open sourced. So about 24 hours later and it kind of took off in a way that I never would have expected. Like I thought it'd be this cool utility that like hardcore stable diffusion users would find useful. But it turns out that almost anyone who mentioned stable diffusion would also kind of mention Lexica in conjunction with it. I think it's because it was like it captured the zeitgeist in an easy to share way where it's like this URL and there's this gallery and you can search. Whereas running the model locally was a lot harder. You'd have to like to deploy it on your own GPU and like set up your own environment and like do all that stuff. Swyx: Oh, my takeaway. I have two more to add to the reasons why Lexica works at the time. One is lower latency is all you need. So in other words, instead of waiting a minute for your image, you could just search and find stuff that other people have done. That's good. And then two is everyone knew how to search already, but people didn't know how to prompt. So you were the bridge. Sharif: That's true. Yeah. You would get a lot better looking images by typing a one word prompt versus prompting for that one word. Yeah. Swyx: Yeah. That is interesting. [11:00] Lexica's Explosion at LaunchAlessio: The numbers kind of speak for themselves, right? Like 24 hours post launch, 51,000 queries, like 2.2 terabytes in bandwidth. Going back to the bandwidth problem that you have before, like you would have definitely run into that. Day two, you doubled that. It's like 111,000 queries, four and a half terabytes in bandwidth, 22 million images served. So it's pretty crazy. Sharif: Yeah. I think we're, we're doing like over 5 billion images served per month now. It's like, yeah, that's, it's pretty crazy how much things have changed since then. Swyx: Yeah. I'm still showing people like today, even today, you know, it's been a few months now. This is where you start to learn image prompting because they don't know. Sharif: Yeah, it is interesting. And I, it's weird because I didn't really think it would be a company. I thought it would just be like a cool utility or like a cool tool that I would use for myself. And I really was just building it for myself just because I didn't want to use the Discord search bar. But yeah, it was interesting that a lot of other people found it pretty useful as well. [11:00] How Lexica WorksSwyx: So there's a lot of things that you release in a short amount of time. The God mode search was kind of like, obviously the first thing, I guess, like maybe to talk about some of the underlying technology you're using clip to kind of find, you know, go from image to like description and then let people search it. Maybe talk a little bit about what it takes to actually make the search magic happen. Sharif: Yeah. So the original search was just using Postgres' full text search and it would only search the text contents of the prompt. But I was inspired by another website called Same Energy, where like a visual search engine. It's really cool. Do you know what happened to that guy? I don't. Swyx: He released it and then he disappeared from the internet. Sharif: I don't know what happened to him, but I'm sure he's working on something really cool. He also worked on like Tabnine, which was like the very first version of Copilot or like even before Copilot was Copilot. But yeah, inspired by that, I thought like being able to search images by their semantics. The contents of the image was really interesting. So I pretty much decided to create a search index on the clip embeddings, the clip image embeddings of all the images. And when you would search it, we would just do KNN search on pretty much the image embedding index. I mean, we had way too many embeddings to store on like a regular database. So we had to end up using FAISS, which is a Facebook library for really fast KNN search and embedding search. That was pretty fun to set up. It actually runs only on CPUs, which is really cool. It's super efficient. You compute the embeddings on GPUs, but like you can serve it all on like an eight core server and it's really, really fast. Once we released the semantic search on the clip embeddings, people were using the search way more. And you could do other cool things. You could do like similar image search where if you found like a specific image you liked, you could upload it and it would show you relevant images as well. Swyx: And then right after that, you raised your seed money from AI grant, NetFreedman, then Gross. Sharif: Yeah, we raised about $5 million from Daniel Gross. And then we also participated in AI grant. That was pretty cool. That was kind of the inflection point. Not much before that point, Lexic was kind of still a side project. And I told myself that I would focus on it full time or I'd consider focusing on it full time if we had broke like a million users. I was like, oh, that's gonna be like years away for sure. And then we ended up doing that in like the first week and a half. I was like, okay, there's something here. And it was kind of that like deal was like growing like pretty slowly and like pretty linearly. And then Lexica was just like this thing that just kept going up and up and up. And I was so confused. I was like, man, people really like looking at pictures. This is crazy. Yeah. And then we decided to pivot the entire company and just focus on Lexica full time at that point. And then we raised our seed round. [15:00] Being Chronically EarlySwyx: Yeah. So one thing that you casually dropped out, the one that slip, you said you were working on Lexica before the launch of Stable Diffusion such that you were able to launch Lexica one day after Stable Diffusion. Sharif: Yeah.Swyx: How did you get so early into Stable Diffusion? Cause I didn't hear about it. Sharif: Oh, that's a good question. I, where did I first hear about Stable Diffusion? I'm not entirely sure. It must've been like somewhere on Twitter or something. That changed your life. Yeah, it was great. And I got into the discord cause I'd used Dolly too before, but, um, there were a lot of restrictions in place where you can generate human faces at the time. You can do that now. But when I first got access to it, like you couldn't do any faces. It was like, there were like a, the list of adjectives you couldn't use was quite long. Like I had a friend from Pakistan and it can generate anything with the word Pakistan in it for some reason. But Stable Diffusion was like kind of the exact opposite where there were like very, very few rules. So that was really, really fun and interesting, especially seeing the chaos of like a bunch of other people also using it right in front of you. That was just so much fun. And I just wanted to do something with it. I thought it was honestly really fun. Swyx: Oh, well, I was just trying to get tips on how to be early on things. Cause you're pretty consistently early to things, right? You were Stadia before Stadia. Um, and then obviously you were on. Sharif: Well, Stadia is kind of shut down now. So I don't know if being early to that was a good one. Swyx: Um, I think like, you know, just being consistently early to things that, uh, you know, have a lot of potential, like one of them is going to work out and you know, then that's how you got Lexica. [16:00] From Search to Custom ModelsAlessio: How did you decide to go from search to running your own models for a generation? Sharif: That's a good question. So we kind of realized that the way people were using Lexica was they would have Lexica open in one tab and then in another tab, they'd have a Stable Diffusion interface. It would be like either a discord or like a local run interface, like the automatic radio UI, um, or something else. I just, I would watch people use it and they would like all tabs back and forth between Lexica and their other UI. And they would like to scroll through Lexica, click on the prompt, click on an image, copy the prompt, and then paste it and maybe change a word or two. And I was like, this should really kind of just be all within Lexica. Like, it'd be so cool if you could just click a button in Lexica and get an editor and generate your images. And I found myself also doing the all tab thing, or it was really frustrating. I was like, man, this is kind of tedious. Like I really wish it was much simpler. So we just built generations directly within Lexica. Um, so we do, we deployed it on, I don't remember when we first launched, I think it was November, December. And yeah, people love generating directly within it. [17:00] AI Grant LearningsSwyx: I was also thinking that this was coming out of AI grants where, you know, I think, um, yeah, I was like a very special program. I was just wondering if you learned anything from, you know, that special week where everyone was in town. Sharif: Yeah, that was a great week. I loved it. Swyx: Yeah. Bring us, bring us in a little bit. Cause it was awesome. There. Sharif: Oh, sure. Yeah. It's really, really cool. Like all the founders in AI grants are like fantastic people. And so I think the main takeaway from the AI grant was like, you have this massive overhang in compute or in capabilities in terms of like these latest AI models, but to the average person, there's really not that many products that are that cool or useful to them. Like the latest one that has hit the zeitgeist was chat GPT, which used arguably the same GPT three model, but like RLHF, but you could have arguably built like a decent chat GPT product just using the original GPT three model. But no one really did it. Now there were some restrictions in place and opening. I like to slowly release them over the few months or years after they release the original API. But the core premise behind AI grants is that there are way more capabilities than there are products. So focus on building really compelling products and get people to use them. And like to focus less on things like hitting state of the art on evals and more on getting users to use something. Swyx: Make something people want.Sharif: Exactly. Host: Yeah, we did an episode on LLM benchmarks and we kind of talked about how the benchmarks kind of constrain what people work on, because if your model is not going to do well, unlike the well-known benchmarks, it's not going to get as much interest and like funding. So going at it from a product lens is cool. [19:30] The Text to Image Illuminati?Swyx: My hypothesis when I was seeing the sequence of events for AI grants and then for Lexica Aperture was that you had some kind of magical dinner with Emad and David Holtz. And then they taught you the secrets of training your own model. Is that how it happens? Sharif: No, there's no secret dinner. The Illuminati of text to image. We did not have a meeting. I mean, even if we did, I wouldn't tell you. But it really boils down to just having good data. If you think about diffusion models, really the only thing they do is learn a distribution of data. So if you have high quality data, learn that high quality distribution. Or if you have low quality data, it will learn to generate images that look like they're from that distribution. So really it boils down to the data and the amount of data you have and that quality of that data, which means a lot of the work in training high quality models, at least diffusion models, is not really in the model architecture, but rather just filtering the data in a way that makes sense. So for Lexica, we do a lot of aesthetic scoring on images and we use the rankings we get from our website because we get tens of millions of people visiting it every month. So we can capture a lot of rankings. Oh, this person liked this image when they saw this one right next to it. Therefore, they probably preferred this one over that. You can do pairwise ranking to rank images and then compute like ELO scores. You can also just train aesthetic models to learn to classify a model, whether or not someone will like it or whether or not it's like, rank it on a scale of like one to ten, for example. So we mostly use a lot of the traffic we get from Lexica and use that to kind of filter our data sets and use that to train better aesthetic models. [20:30] How to Learn to Train ModelsSwyx: You had been a machine learning engineer before. You've been more of an infrastructure guy. To build, you were more of a prompt engineer with a bit of web design. This was the first time that you were basically training your own model. What was the wrap up like? You know, not to give away any secret sauce, but I think a lot of people who are traditional software engineers are feeling a lot of, I don't know, fear when encountering these kinds of domains. Sharif: Yeah, I think it makes a lot of sense. And to be fair, I didn't have much experience training massive models at this scale before I did it. A lot of times it's really just like, in the same way when you're first learning to program, you would just take the problem you're having, Google it, and go through the stack overflow post. And then you figure it out, but ultimately you will get to the answer. It might take you a lot longer than someone who's experienced, but I think there are enough resources out there where it's possible to learn how to do these things. Either just reading through GitHub issues for relevant models. Swyx: Oh God. Sharif: Yeah. It's really just like, you might be slower, but it's definitely still possible. And there are really great courses out there. The Fast AI course is fantastic. There's the deep learning book, which is great for fundamentals. And then Andrej Karpathy's online courses are also excellent, especially for language modeling. You might be a bit slower for the first few months, but ultimately I think if you have the programming skills, you'll catch up pretty quickly. It's not like this magical dark science that only three people in the world know how to do well. Probably was like 10 years ago, but now it's becoming much more open. You have open source collectives like Eleuther and LAION, where they like to share the details of their large scale training runs. So you can learn from a lot of those people. Swyx: Yeah. I think what is different for programmers is having to estimate significant costs upfront before they hit run. Because it's not a thing that you normally consider when you're coding, but yeah, like burning through your credits is a fear that people have. Sharif: Yeah, that does make sense. In that case, like fine tuning larger models gets you really, really far. Even using things like low rank adaptation to fine tune, where you can like fine tune much more efficiently on a single GPU. Yeah, I think people are underestimating how far you can really get just using open source models. I mean, before Lexica, I was working on Debuild and we were using the GP3 API, but I was also like really impressed at how well you could get open source models to run by just like using the API, collecting enough samples from like real world user feedback or real world user data using your product. And then just fine tuning the smaller open source models on those examples. And now you have a model that's pretty much state of the art for your specific domain. Whereas the runtime cost is like 10 times or even 100 times cheaper than using an API. Swyx: And was that like GPT-J or are you talking BERT? Sharif: I remember we tried GPT-J, but I think FLAN-T5 was like the best model we were able to use for that use case. FLAN-T5 is awesome. If you can, like if your prompt is small enough, it's pretty great. And I'm sure there are much better open source models now. Like Vicuna, which is like the GPT-4 variant of like Lama fine tuned on like GPT-4 outputs. Yeah, they're just going to get better and they're going to get better much, much faster. Swyx: Yeah. We're just talking in a previous episode to the creator of Dolly, Mike Conover, which is actually commercially usable instead of Vicuna, which is a research project. Sharif: Oh, wow. Yeah, that's pretty cool. [24:00] Why No Agents?Alessio: I know you mentioned being early. Obviously, agents are one of the hot things here. In 2021, you had this, please buy me AirPods, like a demo that you tweeted with the GPT-3 API. Obviously, one of the things about being early in this space, you can only do one thing at a time, right? And you had one tweet recently where you said you hoped that that demo would open Pandora's box for a bunch of weird GPT agents. But all we got were docs powered by GPT. Can you maybe talk a little bit about, you know, things that you wish you would see or, you know, in the last few, last few weeks, we've had, you know, Hugging GPT, Baby AGI, Auto GPT, all these different kind of like agent projects that maybe now are getting closer to the, what did you say, 50% of internet traffic being skips of GPT agents. What are you most excited about, about these projects and what's coming? Sharif: Yeah, so we wanted a way for users to be able to paste in a link for the documentation page for a specific API, and then describe how to call that API. And then the way we would need to pretty much do that for Debuild was we wondered if we could get an agent to browse the docs page, read through it, summarize it, and then maybe even do things like create an API key and register it for that user. To do that, we needed a way for the agent to read the web page and interact with it. So I spent about a day working on that demo where we just took the web page, serialized it into a more compact form that fit within the 2048 token limit of like GPT-3 at the time. And then just decide what action to do. And then it would, if the page was too long, it would break it down into chunks. And then you would have like a sub prompt, decide on which chunk had the best action. And then at the top node, you would just pretty much take that action and then run it in a loop. It was really, really expensive. I think that one 60 second demo cost like a hundred bucks or something, but it was wildly impractical. But you could clearly see that agents were going to be a thing, especially ones that could read and write and take actions on the internet. It was just prohibitively expensive at the time. And the context limit was way too small. But yeah, I think it seems like a lot of people are taking it more seriously now, mostly because GPT-4 is way more capable. The context limit's like four times larger at 8,000 tokens, soon 32,000. And I think the only problem that's left to solve is finding a really good representation for a webpage that allows it to be consumed by a text only model. So some examples are like, you could just take all the text and pass it in, but that's probably too long. You could take all the interactive only elements like buttons and inputs, but then you miss a lot of the relevant context. There are some interesting examples, which I really like is you could run the webpage or you could run the browser in a terminal based browser. So there are some browsers that run in your terminal, which serialize everything into text. And what you can do is just take that frame from that terminal based browser and pass that directly to the model. And it's like a really, really good representation of the webpage because they do things where for graphical elements, they kind of render it using ASCII blocks. But for text, they render it as actual text. So you could just remove all the weird graphical elements, just keep all the text. And that works surprisingly well. And then there are other problems to solve, which is how do you get the model to take an action? So for example, if you have a booking page and there's like a calendar and there are 30 days on the calendar, how do you get it to specify which button to press? It could say 30, and you can match string based and like find the 30. But for example, what if it's like a list of friends in Facebook and trying to delete a friend? There might be like 30 delete buttons. How do you specify which one to click on? The model might say like, oh, click on the one for like Mark. But then you'd have to figure out the delete button in relation to Mark. And there are some ways to solve this. One is there's a cool Chrome extension called Vimium, which lets you use Vim in your Chrome browser. And what you do is you can press F and over every interactive element, it gives you like a character or two characters. Or if you type those two characters, it presses that button or it opens or focuses on that input. So you could combine a lot of these ideas and then get a really good representation of the web browser in text, and then also give the model a really, really good way to control the browser as well. And I think those two are the core part of the problem. The reasoning ability is definitely there. If a model can score in the top 10% on the bar exam, it can definitely browse a web page. It's really just how do you represent text to the model and how do you get the model to perform actions back on the web page? Really, it's just an engineering problem. Swyx: I have one doubt, which I'd love your thoughts on. How do you get the model to pause when it doesn't have enough information and ask you for additional information because you under specified your original request? Sharif: This is interesting. I think the only way to do this is to have a corpus where your training data is like these sessions of agents browsing the web. And you have to pretty much figure out where the ones that went wrong or the agents that went wrong, or did they go wrong and just replace it with, hey, I need some help. And then if you were to fine tune a larger model on that data set, you would pretty much get them to say, hey, I need help on the instances where they didn't know what to do next. Or if you're using a closed source model like GPT-4, you could probably tell it if you're uncertain about what to do next, ask the user for help. And it probably would be pretty good at that. I've had to write a lot of integration tests in my engineering days and like the dome. Alessio: They might be over. Yeah, I hope so. I hope so. I don't want to, I don't want to deal with that anymore. I, yeah, I don't want to write them the old way. Yeah. But I'm just thinking like, you know, we had the robots, the TXT for like crawlers. Like I can definitely see the DOM being reshaped a little bit in terms of accessibility. Like sometimes you have to write expats that are like so long just to get to a button. Like there should be a better way to do it. And maybe this will drive the change, you know, making it easier for these models to interact with your website. Sharif: There is the Chrome accessibility tree, which is used by screen readers, but a lot of times it's missing a lot of, a lot of useful information. But like in a perfect world, everything would be perfectly annotated for screen readers and we could just use that. That's not the case. [29:30] GPT4 and MultimodalitySwyx: GPT-4 multimodal, has your buddy, Greg, and do you think that that would solve essentially browser agents or desktop agents? Sharif: Greg has not come through yet, unfortunately. But it would make things a lot easier, especially for graphically heavy web pages. So for example, you were using Yelp and like using the map view, it would make a lot of sense to use something like that versus a text based input. Where, how do you serialize a map into text? It's kind of hard to do that. So for more complex web pages, that would make it a lot easier. You get a lot more context to the model. I mean, it seems like that multimodal input is very dense in the sense that it can read text and it can read it really, really well. So you could probably give it like a PDF and it would be able to extract all the text and summarize it. So if it can do that, it could probably do anything on any webpage. Swyx: Yeah. And given that you have some experience integrating Clip with language models, how would you describe how different GPT-4 is compared to that stuff? Sharif: Yeah. Clip is entirely different in the sense that it's really just good at putting images and text into the same latent space. And really the only thing that's useful for is similarity and clustering. Swyx: Like literally the same energy, right? Sharif: Yeah. Swyx: Yeah. And then there's Blip and Blip2. I don't know if you like those. Sharif: Yeah. Blip2 is a lot better. There's actually a new project called, I think, Mini GPT-4. Swyx: Yes. It was just out today. Sharif: Oh, nice. Yeah. It's really cool. It's actually really good. I think that one is based on the Lama model, but yeah, that's, that's like another. Host: It's Blip plus Lama, right? So they, they're like running through Blip and then have Lama ask your, interpret your questions so that you do visual QA. Sharif: Oh, that's cool. That's really clever. Yeah. Ensemble models are really useful. Host: Well, so I was trying to articulate, cause that was, that's, there's two things people are talking about today. You have to like, you know, the moment you wake up, you open Hacker News and go like, all right, what's, what's the new thing today? One is Red Pajama. And then the other one is Mini GPT-4. So I was trying to articulate like, why is this not GPT-4? Like what is missing? And my only conclusion was it just doesn't do OCR yet. But I wonder if there's anything core to this concept of multimodality that you have to train these things together. Like what does one model doing all these things do that is separate from an ensemble of models that you just kind of duct tape together? Sharif: It's a good question. This is pretty related to interoperability. Like how do we understand that? Or how, how do we, why do models trained on different modalities within the same model perform better than two models perform or train separately? I can kind of see why that is the case. Like, it's kind of hard to articulate, but when you have two different models, you get the reasoning abilities of a language model, but also like the text or the vision understanding of something like Clip. Whereas Clip clearly lacks the reasoning abilities, but if you could somehow just put them both in the same model, you get the best of both worlds. There were even cases where I think the vision version of GPT-4 scored higher on some tests than the text only version. So like there might even be some additional learning from images as well. Swyx: Oh yeah. Well, uh, the easy answer for that was there was some chart in the test. That wasn't translated. Oh, when I read that, I was like, Oh yeah. Okay. That makes sense. Sharif: That makes sense. I thought it'd just be like, it sees more of the world. Therefore it has more tokens. Swyx: So my equivalent of this is I think it's a well-known fact that adding code to a language model training corpus increases its ability to do language, not just with code. So, the diversity of datasets that represent some kind of internal logic and code is obviously very internally logically consistent, helps the language model learn some internal structure. Which I think, so, you know, my ultimate test for GPT-4 is to show the image of like, you know, is this a pipe and ask it if it's a pipe or not and see what it does. Sharif: Interesting. That is pretty cool. Yeah. Or just give it a screenshot of your like VS code editor and ask it to fix the bug. Yeah. That'd be pretty wild if it could do that. Swyx: That would be adult AGI. That would be, that would be the grownup form of AGI. [33:30] Sharif's Startup ManualSwyx: On your website, you have this, um, startup manual where you give a bunch of advice. This is fun. One of them was that you should be shipping to production like every two days, every other day. This seems like a great time to do it because things change every other day. But maybe, yeah, tell some of our listeners a little bit more about how you got to some of these heuristics and you obviously build different projects and you iterate it on a lot of things. Yeah. Do you want to reference this? Sharif: Um, sure. Yeah, I'll take a look at it. Swyx: And we'll put this in the show notes, but I just wanted you to have the opportunity to riff on this, this list, because I think it's a very good list. And what, which one of them helped you for Lexica, if there's anything, anything interesting. Sharif: So this list is, it's pretty funny. It's mostly just like me yelling at myself based on all the mistakes I've made in the past and me trying to not make them again. Yeah. Yeah. So I, the first one is like, I think the most important one is like, try when you're building a product, try to build the smallest possible version. And I mean, for Lexica, it was literally a, literally one screen in the react app where a post-process database, and it just showed you like images. And I don't even know if the first version had search. Like I think it did, but I'm not sure. Like, I think it was really just like a grid of images that were randomized, but yeah, don't build the absolute smallest thing that can be considered a useful application and ship it for Lexica. That was, it helps me write better prompts. That's pretty useful. It's not that useful, but it's good enough. Don't fall into the trap of intellectual indulgence with over-engineering. I think that's a pretty important one for myself. And also anyone working on new things, there's often times you fall into the trap of like thinking you need to add more and more things when in reality, like the moment it's useful, you should probably get in the hands of your users and they'll kind of set the roadmap for you. I know this has been said millions of times prior, but just, I think it's really, really important. And I think if I'd spent like two months working on Lexica, adding a bunch of features, it wouldn't have been anywhere as popular as it was if I had just released the really, really boiled down version alongside the stable diffusion release. Yeah. And then there are a few more like product development doesn't start until you launch. Think of your initial product as a means to get your users to talk to you. It's also related to the first point where you really just want people using something as quickly as you can get that to happen. And then a few more are pretty interesting. Create a product people love before you focus on growth. If your users are spontaneously telling other people to use your product, then you've built something people love. Swyx: So this is pretty, it sounds like you've internalized Paul Graham's stuff a lot. Yeah. Because I think he said stuff like that. Sharif: A lot of these are just probably me taking notes from books I found really interesting or like PG essays that were really relevant at the time. And then just trying to not forget them. I should probably read this list again. There's some pretty personalized advice for me here. Oh yeah. One of my favorite ones is, um, don't worry if what you're building doesn't sound like a business. Nobody thought Facebook would be a $500 billion company. It's easy to come up with a business model. Once you've made something people want, you can even make pretty web forms and turn that into a 200 person company. And then if you click the link, it's to LinkedIn for type form, which is now, uh, I think they're like an 800 person company or something like that. So they've grown quite a bit. There you go. Yeah. Pretty web forms are pretty good business, even though it doesn't sound like it. Yeah. It's worth a billion dollars. [38:30] Lexica Aperture V1/2/3Swyx: One way I would like to tie that to the history of Lexica, which we didn't go over, which was just walk us through like Aperture V1, V2, V3, uh, which you just released last week. And how maybe some of those principles helped you in that journey.Sharif: Yeah. So, um, V1 was us trying to create a very photorealistic version of our model of Sable to Fusion. Uh, V1 actually didn't turn out to be that popular. It turns out people loved not generating. Your marketing tweets were popular. They were quite popular. So I think at the time you couldn't get Sable to Fusion to generate like photorealistic images that were consistent with your prompt that well. It was more so like you were sampling from this distribution of images and you could slightly pick where you sampled from using your prompt. This was mostly just because the clip text encoder is not the best text encoder. If you use a real language model, like T5, you get much better results. Like the T5 XXL model is like a hundred times larger than the clip text encoder for Sable to Fusion 1.5. So you could kind of steer it into like the general direction, but for more complex prompts, it just didn't work. So a lot of our users actually complained that they preferred the 1.5, Sable to Fusion 1.5 model over the Aperture model. And it was just because a lot of people were using it to create like parts and like really weird abstract looking pictures that didn't really work well with the photorealistic model trained solely on images. And then for V2, we kind of took that into consideration and then just trained it more on a lot of the art images on Lexica. So we took a lot of images that were on Lexica that were art, used that to train aesthetic models that ranked art really well, and then filtered larger sets to train V2. And then V3 is kind of just like an improved version of that with much more data. I'm really glad we didn't spend too much time on V1. I think we spent about one month working on it, which is a lot of time, but a lot of the things we learned were useful for training future versions. Swyx: How do you version them? Like where do you decide, okay, this is V2, this is V3? Sharif: The versions are kind of weird where you can't really use semantic versions because like if you have a small update, you usually just make that like V2. Versions are kind of used for different base models, I'd say. So if you have each of the versions were a different base model, but we've done like fine tunes of the same version and then just release an update without incrementing the version. But I think when there's like a clear change between running the same prompt on a model and you get a different image, that should probably be a different version. [40:00] Request for AI Startup - LLM ToolsAlessio: So the startup manual was the more you can actually do these things today to make it better. And then you have a whole future page that has tips from, you know, what the series successor is going to be like to like why everyone's genome should be sequenced. There's a lot of cool stuff in there. Why do we need to develop stimulants with shorter half-lives so that we can sleep better. Maybe talk a bit about, you know, when you're a founder, you need to be focused, right? So sometimes there's a lot of things you cannot build. And I feel like this page is a bit of a collection of these. Like, yeah. Are there any of these things that you're like, if I were not building Lexica today, this is like a very interesting thing. Sharif: Oh man. Yeah. There's a ton of things that I want to build. I mean, off the top of my head, the most exciting one would be better tools for language models. And I mean, not tools that help us use language models, but rather tools for the language models themselves. So things like giving them access to browsers, giving them access to things like payments and credit cards, giving them access to like credit cards, giving them things like access to like real world robots. So like, it'd be cool if you could have a Boston dynamic spot powered by a language model reasoning module and you would like to do things for you, like go and pick up your order, stuff like that. Entirely autonomously given like high level commands. That'd be like number one thing if I wasn't working on Lexica. [40:00] Sequencing your GenomeAnd then there's some other interesting things like genomics I find really cool. Like there's some pretty cool things you can do with consumer genomics. So you can export your genome from 23andMe as a text file, like literally a text file of your entire genome. And there is another tool called Prometheus, I think, where you upload your 23andMe text file genome and then they kind of map specific SNPs that you have in your genome to studies that have been done on those SNPs. And it tells you really, really useful things about yourself. Like, for example, I have the SNP for this thing called delayed sleep phase disorder, which makes me go to sleep about three hours later than the general population. So like I used to always be a night owl and I never knew why. But after using Prometheus it pretty much tells you, oh, you have the specific genome for specific SNP for DSPS. It's like a really tiny percentage of the population. And it's like something you should probably know about. And there's a bunch of other things. It tells you your likelihood for getting certain diseases, for certain cancers, oftentimes, like even weird personality traits. There's one for like, I have one of the SNPs for increased risk taking and optimism, which is pretty weird. That's an actual thing. Like, I don't know how. This is the founder gene. You should sequence everybody. It's pretty cool. And it's like, it's like $10 for Prometheus and like 70 bucks for 23andMe. And it explains to you how your body works and like the things that are different from you or different from the general population. Wow. Highly recommend everyone do it. Like if you're, if you're concerned about privacy, just purchase a 23andMe kit with a fake name. You don't have to use your real name. I didn't use my real name. Swyx: It's just my genes. Worst you can do is clone me. It ties in with what you were talking about with, you know, we want the future to be like this. And like people are building uninspired B2B SaaS apps and you and I had an exchange about this. [42:00] Believe in Doing Great ThingsHow can we get more people to believe they can do great things? Sharif: That's a good question. And I like a lot of the things I've been working on with GP3. It has been like trying to solve this by getting people to think about more interesting ideas. I don't really know. I think one is just like the low effort version of this is just putting out really compelling demos and getting people inspired. And then the higher effort version is like actually building the products yourself and getting people to like realize this is even possible in the first place. Like I think the baby AGI project and like the GPT Asian projects on GitHub are like in practice today, they're not super useful, but I think they're doing an excellent job of getting people incredibly inspired for what can be possible with language models as agents. And also the Stanford paper where they had like the mini version of Sims. Yeah. That one was incredible. That was awesome. Swyx: It was adorable. Did you see the part where they invented day drinking? Sharif: Oh, they did? Swyx: Yeah. You're not supposed to go to these bars in the afternoon, but they were like, we're going to go anyway. Nice. Sharif: That's awesome. Yeah. I think we need more stuff like that. That one paper is probably going to inspire a whole bunch of teams to work on stuff similar to that. Swyx: And that's great. I can't wait for NPCs to actually be something that you talk to in a game and, you know, have their own lives and you can check in and, you know, they would have their own personalities as well. Sharif: Yeah. I was so kind of off topic. But I was playing the last of us part two and the NPCs in that game are really, really good. Where if you like, point a gun at them and they'll beg for their life and like, please, I have a family. And like when you kill people in the game, they're like, oh my God, you shot Alice. Like they're just NPCs, but they refer to each other by their names and like they plead for their lives. And this is just using regular conditional rules on NPC behavior. Imagine how much better it'd be if it was like a small GPT-4 agent running in every NPC and they had the agency to make decisions and plead for their lives. And I don't know, you feel way more guilty playing that game. Alessio: I'm scared it's going to be too good. I played a lot of hours of Fallout. So I feel like if the NPCs were a lot better, you would spend a lot more time playing the game. Yeah. [44:30] Lightning RoundLet's jump into lightning round. First question is your favorite AI product. Sharif: Favorite AI product. The one I use the most is probably ChatGPT. The one I'm most excited about is, it's actually a company in AI grants. They're working on a version of VS code. That's like an entirely AI powered cursor, yeah. Cursor where you would like to give it a prompt and like to iterate on your code, not by writing code, but rather by just describing the changes you want to make. And it's tightly integrated into the editor itself. So it's not just another plugin. Swyx: Would you, as a founder of a low code prompting-to-code company that pivoted, would you advise them to explore some things or stay away from some things? Like what's your learning there that you would give to them?Sharif: I would focus on one specific type of code. So if I'm building a local tool, I would try to not focus too much on appealing developers. Whereas if I was building an alternative to VS code, I would focus solely on developers. So in that, I think they're doing a pretty good job focusing on developers. Swyx: Are you using Cursor right now? Sharif: I've used it a bit. I haven't converted fully, but I really want to. Okay. It's getting better really, really fast. Yeah. Um, I can see myself switching over sometime this year if they continue improving it. Swyx: Hot tip for, for ChatGPT, people always say, you know, they love ChatGPT. Biggest upgrade to my life right now is the, I forked a menu bar app I found on GitHub and now I just have it running in a menu bar app and I just do command shift G and it pops it up as a single use thing. And there's no latency because it just always is live. And I just type, type in the thing I want and then it just goes away after I'm done. Sharif: Wow. That's cool. Big upgrade. I'm going to install that. That's cool. Alessio: Second question. What is something you thought would take much longer, but it's already here? Like what, what's your acceleration update? Sharif: Ooh, um, it would take much longer, but it's already here. This is your question. Yeah, I know. I wasn't prepared. Um, so I think it would probably be kind of, I would say text to video. Swyx: Yeah. What's going on with that? Sharif: I think within this year, uh, by the end of this year, we'll have like the jump between like the original DALL-E one to like something like mid journey. Like we're going to see that leap in text to video within the span of this year. Um, it's not already here yet. So I guess the thing that surprised me the most was probably the multi-modality of GPT four in the fact that it can technically see things, which is pretty insane. Swyx: Yeah. Is text to video something that Aperture would be interested in? Sharif: Uh, it's something we're thinking about, but it's still pretty early. Swyx: There was one project with a hand, um, animation with human poses. It was also coming out of Facebook. I thought that was a very nice way to accomplish text to video while having a high degree of control. I forget the name of that project. It was like, I think it was like drawing anything. Swyx: Yeah. It sounds familiar. Well, you already answered a year from now. What will people be most surprised by? Um, and maybe the, uh, the usual requests for startup, you know, what's one thing you will pay for if someone built it? Sharif: One thing I would pay for if someone built it. Um, so many things, honestly, I would probably really like, um, like I really want people to build more, uh, tools for language models, like useful tools, give them access to Chrome. And I want to be able to give it a task. And then just, it goes off and spins up a hundred agents that perform that task. And like, sure. Like 80 of them might fail, but like 20 of them might kind of succeed. That's all you really need. And they're agents. You can spin up thousands of them. It doesn't really matter. Like a lot of large numbers are on your side. So that'd be, I would pay a lot of money for that. Even if it was capable of only doing really basic tasks, like signing up for a SAS tool and booking a call or something. If you could do even more things where it could have handled the email, uh, thread and like get the person on the other end to like do something where like, I don't even have to like book the demo. They just give me access to it. That'd be great. Yeah. More, more. Like really weird language model tools would be really fun.Swyx: Like our chat, GPT plugins, a step in the right direction, or are you envisioning something else? Sharif: I think GPT, chat GPT plugins are great, but they seem to only have right-only access right now. I also want them to have, I want these like theoretical agents to have right access to the world too. So they should be able to perform actions on web browsers, have their own email inbox, and have their own credit card with their own balance. Like take it, send emails to people that might be useful in achieving their goal. Ask them for help. Be able to like sign up and register for accounts on tools and services and be able to like to use graphical user interfaces really, really well. And also like to phone home if they need help. Swyx: You just had virtual employees. You want to give them a Brex card, right? Sharif: I wouldn't be surprised if, a year from now there was Brex GPT or it's like Brex cards for your GPT agents. Swyx: I mean, okay. I'm excited by this. Yeah. Kind of want to build it. Sharif: You should. Yeah. Alessio: Well, just to wrap up, we always have like one big takeaway for people, like, you know, to display on a signboard for everyone to see what is the big message to everybody. Sharif: Yeah. I think the big message to everybody is you might think that a lot of the time the ideas you have have already been done by someone. And that may be the case, but a lot of the time the ideas you have are actually pretty unique and no one's ever tried them before. So if you have weird and interesting ideas, you should actually go out and just do them and make the thing and then share that with the world. Cause I feel like we need more people building weird ideas and less people building like better GPT search for your documentation. Host: There are like 10 of those in the recent OST patch. Well, thank you so much. You've been hugely inspiring and excited to see where Lexica goes next. Sharif: Appreciate it. Thanks for having me. Get full access to Latent Space at www.latent.space/subscribe

Chinchilla Squeaks
Silicon Valley Tank(s)

Chinchilla Squeaks

Play Episode Listen Later Mar 16, 2023 46:50


Hello everyone welcome to my regular ramble through geeky subjects from tech, to history, games, writing, language, and more.This issue is a little light on links, but I have a great interview with Marshall Jung of Tabnine you can hear in the voiceover section of this newsletter or wherever you find your podcasts. We spoke about the company's take on AI coding assistants and how they've been doing things differently from some time.Thanks for reading Chinchilla Squeaks! Subscribe for free to receive new posts and support my work.That aside I look at some of the best content for digging into what happened with Silicon Valley Bank, what will AI to do music, and more!xx ChinchTech"It's always going to be very biased towards the data it's trained on" – Red Hot Chili Peppers and Adele mixing engineer Andrew Scheps shares his views on AI machine learning, and his tips →Three-time Grammy Award-winning mixing engineer Andrew Scheps worked on Adele's 21, RHCP's Stadium Arcadium (both of which is won Grammys for) and has mixed for Black Sabbath, Metallica, Beyoncé, Lady Gaga, Neil Diamond and man, many others.How the Weekend-Long Freakout Over Silicon Valley Bank Ended →Unless you happened to take a digital detox this weekend, you probably witnessed a lot of online commotion regarding tech startups, the banking system, and the institution formerly known as Silicon Valley Bank.The Silicon Valley Bank Contagion Is Just Beginning →When Silicon Valley Bank collapsed on March 10, Garry Tan, president and CEO of startup incubator Y Combinator, called SVB's failure “an extinction level event for startups” that “will set startups and innovation back by 10 years or more.What you need to know about the SVB bank rescue plan →Question and answers about the federal government's plan to stop bank runs.LanguageThe Moral Case Against Equity Language →What's a “justice-involved person”? The Sierra Club's Equity Language Guide discourages using the words stand, Americans, blind, and crazy. The first two fail at inclusion, because not everyone can stand and not everyone living in this country is a citizen.Content from meMy creative writing setupHow do I write my creative works? I tell you how…Technical writing with JetBrains' Writerside and GrazieAnd finallyThanks for reading Chinchilla Squeaks! Subscribe for free to receive new posts and support my work. This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit chinchillasqueaks.substack.com

Noticias Marketing
10: Novedad Diabetes, Newsletter en WhatsApp, problemas para TikTok y Twitter abre su código

Noticias Marketing

Play Episode Listen Later Mar 1, 2023 11:58


10: Novedad Diabetes, Newsletter en WhatsApp, problemas para TikTok y Twitter abre su código Hoy es miércoles 1 de marzo 2023, soy Borja Girón y estás escuchado el podcast Noticias Marketing, comenzamos. Recuerda suscribirte al podcast en Spotify para escuchar todos los episodios y no perderte las noticias por solo 9,99€/mes: https://anchor.fm/borjagiron Y es que nunca había visto tanto movimiento en Internet. IA, Elon con Twitter, prohibiciones de uso de TikTok, Telegram, Podcasts en YouTube… El reloj inteligente de Apple podrá detectar la glucosa en sangre. Novedad diabéticos. Apple patenta un sistema para transcribir un mensaje de texto a nota de voz Poder abrir tu coche con tu iPhone está cada vez más cerca Bizum quiere dar el salto definitivo y permitir el pago presencial Artifact, la app de noticias que mezcla aspectos de TikTok y Google ya está disponible. https://artifact.news/ WhatsApp prepara una nueva herramienta de suscripción y envío de newsletters dentro de la app Canadá se une a EEUU y Europa y bloquea TikTok en los móviles de trabajadores del Gobierno Windows 11 integra el nuevo bot IA de Bing en la barra de tareas donde antes estaba Cortana Spotify lanza la función ‘AI DJ en su bolsillo' para brindar una experiencia de escucha de radio personalizada. Mientras algunos locutores anteriormente han expresado su preocupación sobre posiblemente perder su trabajo a causa de la IA. https://www.skyshowtime.com/es, la nueva plataforma de tv en streaming que compite con Netflix lanzan a mitad de precio (2,99€/mes) La plataforma de podcasts Podimo ha lanzado su última campaña de branding y está presente en el medio exterior, concretamente en puntos y calles emblemáticos de Madrid y Barcelona. Alternativas a ChatGPT: Bing Chat, Tabnine, YouChat, ChatSonic, GitHub Copilot. * La empresa Anthropic está creando un chatbot llamado 'Claude' para competir con ChatGPT y desarrollar sistemas informáticos. * Alphabet (la empresa matriz de Google) decidió invertir 400 millones (370 millones de euros, aproximadamente) para crear un chatbot llamado 'Claude'. POR FIN: EL PIXEL WATCH YA PUEDE DETECTAR CAÍDAS Y AYUDARTE SI TIENES PROBLEMAS Google definitivamente ha activado una función que es muy útil Meta presenta su propia inteligencia artificial: LLaMA, y la abre a los investigadores La propietaria de Facebook reacciona al éxito de ChatGPT y defiende su modelo, “más pequeño” El filtro Bold Glamour incendia TikTok Un algoritmo que rejuvenece y transforma los rostros de los usuarios de la red social desata el debate Ya pasamos casi un cuarto del día ‘escroleando' en Internet Las últimas investigaciones neurocientíficas indican que la dinámica adictiva de las redes alteran la percepción del tiempo Musk continua con las rondas de despidos en Twitter * Más de 200 empleados han sido despedidos (un 10% de la plantilla): directivos, ingenieros, jefes de producto... * El motivo parece ser el descontento de Musk con lo que había convertido Twitter Blue ELON MUSK SEÑALÓ QUE TWITTER PODRÍA LIBERAR SU ALGORITMO BAJO CÓDIGO ABIERTO DURANTE LA PRÓXIMA SEMANA Existen altas probabilidades de que Elon Musk venda Twitter antes de lo esperado SportsLens revela nuevas probabilidades de que Elon Musk venda Twitter casi un año después de su compra por $43 mil millones Recuerda suscribirte en Spotify para escuchar todos los episodios y no perderte las noticias: https://anchor.fm/borjagiron Soy Borja Girón, has escuchado el podcast Noticias Marketing, nos escuchamos en el próximo episodio. Puedes encontrarme en https://borjagiron.comConviértete en un seguidor de este podcast: https://www.spreaker.com/podcast/noticias-marketing--5762806/support.

Moscow Python: подкаст о Python на русском
Copilot для Python-разработчика. Почему AI не изменил разработку?

Moscow Python: подкаст о Python на русском

Play Episode Listen Later Feb 23, 2023 64:11


В гостях у Moscow Python Podcast Никита Соболев, CTO wemake.services и Арсений Сапелкин, тимлид Kaspersky OS. Обсудили в выпуске: опыт использования Copilot этические вопросы использования AI в работе перспективы Copilot, ChatGPT «у программистов будет больше задач или уменьшится количество программистов?» все станут тимлидами или «киберпанк, который мы заслужили» количество кода будет расти, качество - возможно, но неточно AI системы сделают сильных программистов сильнее, а слабые останутся на том же уровне и не захотят учиться их использовать? как пропасть между новичками и опытными программистами расширится Kite, Tabnine vs Copilot плюсы и минусы систем AI в обучении программированию проблема безопасности: как понять, написан код человеком или AI? ChatGPT, как тьютор и лже-эксперт будущее вместе с AI  Ссылки подкаста:  Все выпуски: https://podcast.python.ru Митапы MoscowPython: https://moscowpython.ru Курс Learn Python: https://learn.python.ru/

ExplAInable
יצירה אוטומטית של קוד עם פרופ ערן יהב וד”ר אורי אלון

ExplAInable

Play Episode Listen Later Dec 19, 2022 61:31


בפרק קודם דיברנו על קודקס, מודל השפה שמאחורי Github Co-Pilot בפרק זה, החלטנו להעמיק ולראיין את ערן יהב, חוקר באקדמיה והמוח מאחורי TabNine ואורי אלון, פוסט דוקטורט בתחום שחוקר את הנושא הרבה לפני שזה היה מגניב. נדבר על ההיסטוריה של התחום, גישות מבוססות גרפים, דקדוק ומה הקשר בין שפת התכנות ליכולת החיזוי. כמו כן נדבר על אתגרים מוצריים ב TabNine ועל המעטפת הנדרשת להפיכת מודל למוצר שאנשים משתמשים בו ומרוצים ממנו.

tabnine
Reversim Podcast
449 Bumpers 80

Reversim Podcast

Play Episode Listen Later Oct 11, 2022


[קישור לקובץ mp3]פודקאסט מספר 449 של רברס עם פלטפורמה - באמפרס מספר 80(!) של ספטמבר 2022רן, אלון ודותן עם אוסף של קצרצרים על חדשות טכנולוגיות מעולם פיתוח התוכנה ומה שמסביב מהזמן האחרון - משתדלים להקליט כל חודש אבל הפעם יצא קצת יותר (רבעוני . . . .440 Bumpers 79].-רן - (רן) אני אתחיל בסדרה של אייטמים בנושא של Code Helpers - אז Code Helpers התחיל כאיזושהי נישה של כלים שעוזרים למפתחים לכתוב את הקוד “בצורה אוטומטית”, או “חצי-אוטומטית” - וזה נמצא כבר כמה שנים המפורסמים שבהם זה אולי GitHub של Copilot - שעכשיו נמצא ב-General Availability ואפילו בתשלום.ונזכיר שכבר שנים רבות קיימת חברה ישראלית שנקראית Tabnine שנמצאת בתחום הזה.ולאחרונה השוק הזה די התלהט . . . אז (1) - כמו שאמרנו, זה ש-GitHub Copilot הגיע ל-General Availability ועכשיו הוא למעשה כבר בתשלוםאחרי משהו כמו שלושה חודשי ניסיון, אתם תדרשו לשלם עליו.ומולם קיים Offering בהחלט לא רע גם של החברה הישראלית Tabnineלמעשה, אני מקשר פה לאיזשהו Thread בתוך Twitter של המנכ”ל, דרור, המנכ”ל של Tabnine - שבא ועושה איזושהי סקירה מקיפה של יתרונות וחסרו… קרא עוד

TechCrunch Startups – Spoken Edition
Tabnine raises $15.5M for AI that autocompletes code

TechCrunch Startups – Spoken Edition

Play Episode Listen Later Jun 16, 2022 6:11


Tabnine, a startup creating an “AI-powered assistant” for software developers, today closed a $15.5 million funding round co-led by Qualcomm Ventures, OurCrowd, and Samsung NEXT Ventures with participation from existing backers Khosla Ventures and Headline Ventures.

TechCrunch Startups – Spoken Edition
Tabnine raises $15.5M for AI that autocompletes code

TechCrunch Startups – Spoken Edition

Play Episode Listen Later Jun 16, 2022 6:11


Tabnine, a startup creating an “AI-powered assistant” for software developers, today closed a $15.5 million funding round co-led by Qualcomm Ventures, OurCrowd, and Samsung NEXT Ventures with participation from existing backers Khosla Ventures and Headline Ventures.

Geekonomy - גיקונומי - פודקאסט שבועי על החיים עצמם
פרק #557 - ד״ר ערן יהב עוזר למפתחים

Geekonomy - גיקונומי - פודקאסט שבועי על החיים עצמם

Play Episode Listen Later May 25, 2022 78:01


ערן יהב הוא מדען מחשב ישראלי המתמחה בהנדסת תוכנה. יהב הוא פרופסור חבר בפקולטה למדעי המחשב בטכניון וה-CTO של Tabnine. החברה מסייעת למפתחים ברחבי העולם לכתוב קוד באמצעות השלמות אוטומטיות,  שמבוססות על שימוש נרחב באינטליגנציה מלאכותית.   על מה דיברנו: מיקרוסופט, טאבניין, פיתוח תוכנה, סטאק אוברפלואו, גיטהאב, ליינוס, הטכניון, אקדמיה, שפות תכנות שונות   נותני החסות שלנו: חברת CrowdStrike   קישורים מהפרק: האתר של Tabnine הספר של בן ברננקי עליו ראם דיבר ההרצאה של לינוס על גיט בגוגל  

crowdstrike tabnine
COMPRESSEDfm
55 | VS Code Extensions, Plugins, and Themes (Part 2)

COMPRESSEDfm

Play Episode Listen Later Mar 15, 2022 56:33


This episode is Part 2 of Amy and James's favorite VS Code Hot Tips and Tricks for improving the developer experience. They share their favorite extensions, plugins, and themes for getting the most out of VS Code, including some hot takes on GitHub CoPilot.SponsorsVercelVercel combines the best developer experience with an obsessive focus on end-user performance. Their platform enables frontend teams to do their best work. It is the best place to deploy any frontend app. Start by deploying with zero configuration to their global edge network. Scale dynamically to millions of pages without breaking a sweat.For more information, visit Vercel.comZEAL is hiring!ZEAL is a computer software agency that delivers “the world's most zealous” and custom solutions. The company plans and develops web and mobile applications that consistently help clients draw in customers, foster engagement, scale technologies, and ensure delivery.ZEAL believes that a business is “only as strong as” its team and cares about culture, values, a transparent process, leveling up, giving back, and providing excellent equipment. The company has staffers distributed throughout the United States, and as it continues to grow, ZEAL looks for collaborative, object-oriented, and organized individuals to apply for open roles.For more information visit softwareresidency.com/careersDatoCMSDatoCMS is a complete and performant headless CMS built to offer the best developer experience and user-friendliness in the market. It features a rich, CDN-powered GraphQL API (with realtime updates!), a super-flexible way to handle dynamic layouts and structured content, and best-in-class image/video support, with progressive/LQIP image loading out-of-the-box."For more information, visit datocms.comShow Notes0:00 Introduction6:59 Our Favorite Extensions7:32 Code Snap and Polacode10:57 Better Comments12:14 Bookmarks13:02 Sponsor: DatoCMS13:56 Cloak14:37 Indent 4 to 216:02 CSS Peak16:48 Error Lens17:34 File Utils19:13 Import Cost21:07 Project Manager21:20 Auto Complete22:09 Tabnine and Kite23:07 GitHub Co-Pilot25:19 Sponsor: ZEAL26:12 Git Integration and Git Lens27:23 GitHub Pull Requests and Issues27:44 LiveShare29:04 IntelliSense for CSS Class Names in HTML30:29 Snippets31:58 Adding Extensions33:07 Thunder ClientJames's YouTube Video on Thunder Client35:04 Calculator35:34 Markdown PDF36:09 Sponsor: Vercel37:15 change-case38:10 Prisma38:43 Color Bracket39:34 Quokka.js40:20 Colorize40:50 Text Pastry41:16 Emmet42:00 Window Colors43:34 Peacock43:55 Building Your Own Extensions44:32 Cobalt 245:37 Other ThemesNight OwlWinter is ComingMidnight SynthcodeSTACKrLevel up TutsShades of PurpleHot Dog Stand47:45 Grab Bag Questions47:56 Question #1: How Hard is it to code your own VS Code Extension?48:10 Question #2: Have you heard of Thunder Client?Thunder Client48:25 Question #3: What do you think of Beginner Developers Using Extensions to Make Things Easier?51:20 Question #4: Any References or Guides on Creating a VS Code Extension that You guys have used?52:34 Picks and Plugs52:36 James's Pick: Duolingo App54:03 James's Plug: James Q Quick on YouTube54:31 Amy's Pick: Pacific Northwest Backpack / Arkadia Supply Co55:07 Amy's Plug: Amy's YouTube Channel

Programming Throwdown
127: AI for Code with Eran Yahav

Programming Throwdown

Play Episode Listen Later Feb 14, 2022 68:59


Brief Summary:Programming is difficult as it is, but imagine how difficult it was without all the current tools, compilers, synthesizers, etc. that we have today. Eran Yahav, Chief Technology Officer at Tabnine shares how AI is currently helping with code writing and how it could change in the future.00:00:16 Introduction00:00:51 Eran Yahav's programming background00:08:11 Balance between Human and the Machine00:11:49 Static Analysis00:29:42 Similarities in Programming Constructs00:25:30 Average vs Tailored tooling00:36:19 Machine Learning Quality Metrics 00:38:27 Rollbar00:40:19 Model Training vs Statistic Matching00:50:19 Developers Interacting with their Code in the Future01:00:18 Tabnine01:08:17 FarewellsResources mentioned in this episode:Companies:Tabnine:  Website: https://www.tabnine.com/ Twitter: https://twitter.com/Tabnine_ LinkedIn: https://www.linkedin.com/company/tabnine/ Social Media:Eran Yahav, Chief Technology Officer at Tabnine Twitter: https://twitter.com/yahave LinkedIn: https://www.linkedin.com/in/eranyahav/ Sponsor:Rollbar Website: https://rollbar.com/ Freebies: https://try.rollbar.com/pt/ If you've enjoyed this episode, you can listen to more on Programming Throwdown's website: https://www.programmingthrowdown.com/Reach out to us via email: programmingthrowdown@gmail.comYou can also follow Programming Throwdown on Facebook | Apple Podcasts | Spotify | Player.FM Join the discussion on our DiscordHelp support Programming Throwdown through our Patreon★ Support this podcast on Patreon ★

Syntax - Tasty Web Development Treats

In this episode of Syntax, Scott and Wes review your portfolios and websites including some from Harryxli, Austin Baird, Jacks Portfolio, and more! Linode - Sponsor Whether you're working on a personal project or managing enterprise infrastructure, you deserve simple, affordable, and accessible cloud computing solutions that allow you to take your project to the next level. Simplify your cloud infrastructure with Linode's Linux virtual machines and develop, deploy, and scale your modern applications faster and easier. Get started on Linode today with a $100 in free credit for listeners of Syntax. You can find all the details at linode.com/syntax. Linode has 11 global data centers and provides 24/7/365 human support with no tiers or hand-offs regardless of your plan size. In addition to shared and dedicated compute instances, you can use your $100 in credit on S3-compatible object storage, Managed Kubernetes, and more. Visit linode.com/syntax and click on the “Create Free Account” button to get started. LogRocket - Sponsor LogRocket lets you replay what users do on your site, helping you reproduce bugs and fix issues faster. It's an exception tracker, a session re-player and a performance monitor. Get 14 days free at logrocket.com/syntax. Tabnine - Sponsor Tabnine is your teams' go to AI assistant. Using a variety of machine learning models, Tabnine learns from your team's best practices, and suggests code completions based on your code. It supports over 30 languages and is available in most IDEs. Tabnine's universal models are trained strictly on fully permissive open source code, and can run locally, meaning that your code stays yours. Get the free version at tabnine.com/now or go to tabnine.com/promo/syntax to get 50% off your first 3 months of Tabnine Teams. Show Notes 00:11 Welcome 01:05 This chapter is mid Wes does TikTok 04:28 Harry xli Harryxli 15:00 AustinBaird.software AustinBaird.software 19:45 Sponsor: Linode 21:10 ognjenbostjancic.com ognjenbostjancic.com 28:19 ndo.dev ndo.dev 35:37 Sponsor: LogRocket 37:02 jacksportfolio.com JacksPortfolio.com 41:51 einargudni.com einargudni.com 43:22 Kids advert break 46:05 Back to einargudni.com 48:24 Sponsor: Tabnine 50:41 cyrillappert.ch hslu.cyrillappert.ch 54:30 Sick Picks 59:02 Shameless Plugs ××× SIIIIICK ××× PIIIICKS ××× Scott: The Alpinist (2021) - IMDb Wes: Instant Pot Air Fryer Lid Shameless Plugs Scott: Astro Course - Sign up for the year and save 50%! Wes: All Courses Tweet us your tasty treats Scott's Instagram LevelUpTutorials Instagram Wes' Instagram Wes' Twitter Wes' Facebook Scott's Twitter Make sure to include @SyntaxFM in your tweets

Podcast – Software Engineering Daily
Tabnine with Eran Yahav

Podcast – Software Engineering Daily

Play Episode Listen Later Jan 21, 2022 50:24


Tabnine is an AI assistant that helps software engineers write more efficient code.  It's been trained on a large corpus of source code or can be trained based on your specific codebase.  Either way, the resulting model offers predictive completion of code that can make programmers more productive.  In this episode, I interview Eran Yahav, The post Tabnine with Eran Yahav appeared first on Software Engineering Daily.

Software Engineering Daily
Tabnine with Eran Yahav

Software Engineering Daily

Play Episode Listen Later Jan 21, 2022 43:11


Tabnine is an AI assistant that helps software engineers write more efficient code.  It's been trained on a large corpus of source code or can be trained based on your specific codebase.  Either way, the resulting model offers predictive completion of code that can make programmers more productive.  In this episode, I interview Eran Yahav, The post Tabnine with Eran Yahav appeared first on Software Engineering Daily.

The Changelog
AI-assisted development is here to stay

The Changelog

Play Episode Listen Later Dec 17, 2021 76:40 Transcription Available


We're joined by Eran Yahav — talking about AI assistants for developers. Eran has been working on this problem for more than a decade. We talk about his path to now and how the idea for Tabnine came to life, this AI revolution taking place and the role it will play in developer productivity, and we talk about the elephant in the room - how Tabnine compares to GitHub Copilot, and what they're doing to make Tabnine the AI assistant for every developer regardless of the IDE or editor you choose.

Changelog Master Feed
AI-assisted development is here to stay (The Changelog #472)

Changelog Master Feed

Play Episode Listen Later Dec 17, 2021 76:40 Transcription Available


We're joined by Eran Yahav — talking about AI assistants for developers. Eran has been working on this problem for more than a decade. We talk about his path to now and how the idea for Tabnine came to life, this AI revolution taking place and the role it will play in developer productivity, and we talk about the elephant in the room - how Tabnine compares to GitHub Copilot, and what they're doing to make Tabnine the AI assistant for every developer regardless of the IDE or editor you choose.

Talk Python To Me - Python conversations for passionate developers
#337: Kedro for Maintainable Data Science

Talk Python To Me - Python conversations for passionate developers

Play Episode Listen Later Oct 9, 2021 63:14


Have you heard of Kedro? It's a Python framework for creating reproducible, maintainable and modular data science code. We all know that reproducibility and related topics are important ones in the data science space. The freedom to pop open a notebook and just start exploring is much of the magic. Yet, that free-form style can lead to difficulties in versioning, reproducibility, collaboration, and moving to production. Solving these problems is the goal of Kedro. And we have 3 great guests from the Kedro community here to give us the rundown: Yetunde Dada, Waylon Walker, and Ivan Danov. Links from the show Waylong on Twitter: @_WaylonWalker Yetunda on Twitter: @yetudada Ivan on Twitter: @ivandanov Kedro: kedro.readthedocs.io Kedro on GitHub: github.com Join the Kedro Discord: discord.gg Articles about Kedro by Waylan: waylonwalker.com Kedro spaceflights tutorial: kedro.readthedocs.io “Hello World” on Kedro: kedro.readthedocs.io Kedro Viz: quantumblacklabs.github.io Spaceflights Tutorial video: youtube.com Dynaconf package: dynaconf.com fsspec: Filesystem interfaces for Python: filesystem-spec.readthedocs.io Neovim: neovim.io Watch this episode on YouTube: youtube.com Episode transcripts: talkpython.fm ---------- Stay in touch with us ---------- Subscribe on YouTube (for live streams): youtube.com Follow Talk Python on Twitter: @talkpython Follow Michael on Twitter: @mkennedy Sponsors Tabnine Talk Python Training AssemblyAI

Unruly Software
Episode 128: Github Copilot Took My Job

Unruly Software

Play Episode Listen Later Aug 31, 2021 55:00


No bloated ORM is safe this episode. Zapatos, react-query and other libraries are incredible and have replaced many of our old disgusting tools like Typeorm, Hibernate and Redux. We're interested in everything DX so we've given Github Copilot and Tabnine a shot and discuss how much we absolutely hate AI pair programming. Questions? Comments? Find out more on our site podcast.unrulysoftware.com (https://podcast.unrulysoftware.com). You can join our discord (https://discord.gg/NGP2nWtFJb) to chat about tech anytime directly with the hosts.

20 Minute Leaders
Ep519: Dekel Persi | Co Founder and Managing Partner, TPY Capital

20 Minute Leaders

Play Episode Listen Later Aug 8, 2021 23:38


Dekel Persi is a founding partner at TPY Capital, an early-stage venture capital fund with over $135M under management. TPY Capital partners with both serial and first-time entrepreneurs seeking partners who share similar values and hands-on support. TPY Capital invested in more than 20 companies including AI21, UniPaaS, H2Pro, Qedma, Pasqal, Paramount Data, TabNine, Signals, Seebo and Enzymotec.

כל תכני עושים היסטוריה
האם בינה מלאכותית תחליף אותנו כמפתחים? [עושים תוכנה]

כל תכני עושים היסטוריה

Play Episode Listen Later Apr 11, 2021 45:44


בשנים האחרונות אנחנו עדים להתפתחויות מרתקות בתחום הבינה המלאכותית כשאחת ההשפעות שלה שמעוררת חשש גם בקהילות המפתחים היא מקצועות שמוחלפים על ידיה.אירחנו באולפן את פרופ׳ ערן יהב, כיום CTO בחברת Tabnine לדבר על פרודוקטיביות של מפתחים, לנתח מה קשה בעבודה של מפתחים ומה העתיד צופן לנו באינטראקציה בין בינה מלאכותית וקוד.האזנה נעימה, עמית.https://www.ads.ranlevi.com/2021/04/11/tapuach-osimtochna-ai-coders/

tabnine