Podcasts about Datadog

Monitoring platform for cloud applications

  • 430PODCASTS
  • 1,084EPISODES
  • 41mAVG DURATION
  • 5WEEKLY NEW EPISODES
  • Apr 14, 2025LATEST
Datadog

POPULARITY

20172018201920202021202220232024

Categories



Best podcasts about Datadog

Show all podcasts related to datadog

Latest podcast episodes about Datadog

10KMedia Podcast
Episode 63: Aaron Fischer, Principal Attorney at Next Gen Law

10KMedia Podcast

Play Episode Listen Later Apr 14, 2025 50:34


Adam sits down with Aaron to discuss navigating the startup landscape from a legal perspective, Datadog's rocket-ship growth, and how he's helping early-stage startups today.

NoLimitSecu
Sécurisation de la chaîne d’approvisionnement logicielle

NoLimitSecu

Play Episode Listen Later Apr 6, 2025 37:46


Episode #497 consacré à la sécurisation de la chaîne d'approvisionnement logicielle (software supply chain) Avec Christophe Tafani-Dereeper Références : https://www.datadoghq.com/blog/engineering/secure-publication-of-datadog-agent-integrations-with-tuf-and-in-toto https://github.com/DataDog/guarddoghttps://github.com/DataDog/malicious-software-packages-dataset/https://github.com/DataDog/supply-chain-firewall/https://github.com/sigstore/cosign https://in-toto.io/https://slsa.dev/https://deps.dev/https://www.sigstore.dev/ https://openssf.org/package-analysis/https://openssf.org/projects/scorecard/https://github.com/google/osv-scanner/ The post Sécurisation de la chaîne d'approvisionnement logicielle appeared first on NoLimitSecu.

Category Visionaries
Jan Willem Rombouts, CEO & Founder of Beebop AI: $5.5 Million Raised to Power Grid Orchestration for the Clean Energy Transition

Category Visionaries

Play Episode Listen Later Apr 4, 2025 33:15


Beebop AI is pioneering a new middleware layer for power grid orchestration, securing $5.5 million in funding to help utilities and energy retailers optimize energy consumption and costs. In this episode of Category Visionaries, I sat down with Jan Willem Rombouts, CEO and Founder of Beebop AI, to discuss how his background at Goldman Sachs and experience building his first energy tech company shaped his approach to solving one of the energy transition's biggest challenges: balancing power grids in an increasingly renewable-powered world. Topics Discussed: Jan Willem's journey from Goldman Sachs' trading floor during the financial crisis to energy tech entrepreneurship The painful lessons learned building Restore, which pioneered virtual power plants and was later acquired by Centrica How Beebop AI creates a middleware layer that orchestrates power consumption across customer devices like EVs, solar panels, and heat pumps Why power grid orchestration is critical to making renewable energy both reliable and affordable Beebop's strategic flywheel connecting utilities and device manufacturers The go-to-market strategies that helped Beebop gain traction with major European utilities   GTM Lessons For B2B Founders: Engineer network effects into your go-to-market strategy: Beebop designed a utility-to-OEM flywheel where each new utility customer helps bring device manufacturers onto their platform, creating a powerful network effect. Jan Willem explained: "What we designed was that we would first contract these utilities... our anticipation was that they would be able to engage with these OEMs, with these manufacturers more easily, to essentially invite them to integrate with our platform." This approach turns customers into channel partners who can open doors that would be difficult for a startup to access directly. Break through complex sales cycles with land-and-expand: When selling to utilities and large corporations with notoriously long sales cycles, Beebop starts with a low-cost, high-value initial offering focused on insights and business case validation. Jan Willem noted: "Our initial proposition is very low cost and very high value... we allow them to see what the business case is... to create somewhat of a solid launching pad on which we can then expand and go to actual operationalization." This approach shortens time-to-value and creates internal champions. Focus on customer economics, not just your technology: Despite having complex technology, Beebop leads customer conversations with how their solution impacts key metrics like customer lifetime value, margin, churn, and customer acquisition costs. "Before we have explained anything about how new our software is, where it positions in the technology stack, we just show what kind of awesome products they can build... creating tens of percentages of discounts on their energy bills." Design for global scale from day one: Based on lessons from his first company, Jan Willem deliberately architected Beebop to work with market structures that are universal across regions: "What we did this time... is we chose markets that have a universal footprint and so that look essentially the same whether you're in the UK or you're in Texas or you're in Germany or you're in Sweden." This approach avoids the scaling challenges of having to constantly adapt to different regulatory environments. Bring process to event marketing: Beebop transformed their trade show approach by adopting a disciplined, metrics-driven strategy learned from Datadog's former CMO. Jan Willem shared: "The big learning for me was to be super intentional. If you go to a trade show, be super clear about exactly how many marketing qualified, how many sales qualified leads you want out of it, and then engineer a team with different roles and responsibilities." This systematic approach yields measurable ROI from events that many startups struggle to achieve.   //   Sponsors: Front Lines — We help B2B tech companies launch, manage, and grow podcasts that drive demand, awareness, and thought leadership. www.FrontLines.io The Global Talent Co. — We help tech startups find, vet, hire, pay, and retain amazing marketing talent that costs 50-70% less than the US & Europe.  www.GlobalTalent.co

Datacenter Technical Deep Dives
How To Learn Like A Rockstar!

Datacenter Technical Deep Dives

Play Episode Listen Later Apr 3, 2025


Amanda Ruzza is a DevOps Engineer, world famous Jass Bassist, and a Services Architect at Datadog! in this episode she shares how she ‘migrated' traditional music studying techniques into learning Cloud and all things tech related! "Study is fun and it's all about falling in love with the journey

Hunters and Unicorns
How Dan Fougere Scaled 4 Billion-Dollar IPOs (And What He'd Do Differently)

Hunters and Unicorns

Play Episode Listen Later Apr 2, 2025 57:15


In this episode of THE PLAYBOOK UNIVERSE, we sit down with legendary CRO Dan Fougere, a pivotal force behind FOUR $1B IPOs. From engineering roots to building elite go-to-market engines, Dan unpacks the battle-tested frameworks that turned great ideas into generational companies. He shares hard-won lessons from PTC, BladeLogic, Medallia, and Datadog—including how to scale a sales org, recruit world-class talent, and align GTM with visionary founders. Whether you're a startup founder, CRO, or early-stage sales leader, this is a masterclass you don't want to miss. Dan also reveals the make-or-break mindset shifts that helped him navigate adversity, imposter syndrome, and the constant pressure of building under extreme constraints.   This episode is packed with insights you simply won't find in any sales playbook.

PodRocket - A web development podcast from LogRocket
Debugging apps with Deno and OpenTelemetry with Luca Casonato

PodRocket - A web development podcast from LogRocket

Play Episode Listen Later Mar 27, 2025 24:55


Luca Casanato, member of the Deno core team, delves into the intricacies of debugging applications using Deno and OpenTelemetry. Discover how Deno's native integration with OpenTelemetry enhances application performance monitoring, simplifies instrumentation compared to Node.js, and unlocks new insights for developers! Links https://lcas.dev https://x.com/lcasdev https://github.com/lucacasonato https://mastodon.social/@lcasdev https://www.linkedin.com/in/luca-casonato-15946b156 We want to hear from you! How did you find us? Did you see us on Twitter? In a newsletter? Or maybe we were recommended by a friend? Let us know by sending an email to our producer, Emily, at emily.kochanekketner@logrocket.com (mailto:emily.kochanekketner@logrocket.com), or tweet at us at PodRocketPod (https://twitter.com/PodRocketpod). Follow us. Get free stickers. Follow us on Apple Podcasts, fill out this form (https://podrocket.logrocket.com/get-podrocket-stickers), and we'll send you free PodRocket stickers! What does LogRocket do? LogRocket provides AI-first session replay and analytics that surfaces the UX and technical issues impacting user experiences. Start understand where your users are struggling by trying it for free at [LogRocket.com]. Try LogRocket for free today.(https://logrocket.com/signup/?pdr) Special Guest: Luca Casonato.

She Said Privacy/He Said Security
How AI Is Revolutionizing Contract Reviews for Legal Teams

She Said Privacy/He Said Security

Play Episode Listen Later Mar 27, 2025 33:00


Farah Gasmi is the Co-founder and CPO of Dioptra, the accurate and customizable AI agent that drafts playbooks and consistently redlines contracts in Microsoft Word. Dioptra is trusted by some of the most innovative teams, like Y Combinator and Wilson Sonsini. She has over 10 years of experience building AI products in healthcare, insurance, and tech for companies like Spotify. Farah is also an adjunct professor at Columbia Business School in NYC. She teaches a Product Management course with a focus on AI and data products. Laurie Ehrlich is the Chief Legal Officer at Dioptra, a cutting-edge legal tech startup revolutionizing contract redlining and playbook generation with AI. With a background leading legal operations and commercial contracting at Datadog and Cognizant, Laurie has deep expertise in scaling legal functions to drive business impact. She began her career in intellectual property law at top firms and holds a JD from NYU School of Law and a BS from Cornell. Passionate about innovation and diversity in tech, Laurie has also been a champion for women in leadership throughout her career. In this episode… Contract review can be time-consuming and complex, especially when working with third-party agreements that use unfamiliar language and formats. Legal teams often rely on manual review processes that make it challenging to maintain consistency across contracts, contributing to inefficiencies and increased costs. That's why businesses need an effective solution that reduces the burden of contract analysis while supporting legal and strategic decision-making. Dioptra, a legal tech startup, helps solve these challenges by leveraging AI to automate first-pass contract reviews, redline contracts, and generate playbooks. The AI agent analyzes past agreements to identify patterns, standard language, and key risk areas, allowing teams to streamline the review process. It supports a range of use cases — from NDAs to real estate deals — while improving consistency and reducing review time. Dioptra also enhances post-execution analysis by enabling companies to assess past agreements for compliance and risk exposure. In this episode of She Said Privacy/He Said Security, Jodi and Justin Daniels speak with Farah Gasmi, Co-founder and Chief Product Officer at Dioptra, and Laurie Ehrlich, the Chief Legal Officer at Dioptra, about how AI is used to streamline contract reviews. Together, they discuss how Dioptra accelerates contract reviews, supports security and privacy through strict data controls, and enables organizations to build smarter, more consistent contract processes — without removing the need for expert human judgment. Farah and Laurie also delve into the importance of AI-driven consistency in contract negotiation, vendor security evaluations, and how companies can safeguard sensitive data when using AI tools.

The Fintech Blueprint
How Metronome is Building the Revenue Engine for AI, with CEO Scott Woody

The Fintech Blueprint

Play Episode Listen Later Mar 25, 2025 49:32


Lex interviews Scott Woody, CEO and founder of Metronome, a usage-based billing platform. Scott shares his journey from academia to entrepreneurship, detailing his experiences at UC Berkeley, D.E. Shaw, and Stanford, where he studied biophysics. His tenure at Dropbox, where he tackled billing system challenges, inspired the creation of Metronome. The discussion highlights Metronome's real-time billing data capabilities, which aim to improve business efficiency and customer experience. Scott also explores the broader implications of AI in fintech, emphasizing the shift towards usage-based business models and the importance of real-time data. Notable discussion points: Metronome emerged from firsthand frustrations at Dropbox, where Scott Woody experienced how rigid billing systems slowed growth, confused customers, and blocked real-time insights. He built Metronome as a flexible, real-time billing engine that merges usage data with pricing logic—powering the monetization infrastructure for top AI companies today. Real-time billing isn't just about invoices—it's a strategic revenue lever. For AI and SaaS businesses alike, Metronome enables teams to run dynamic experiments, optimize GPU allocation, and make last-minute decisions to hit quarterly targets—turning billing into a core growth engine. The rise of AI is accelerating a shift to usage-based models. As AI becomes specialized labor across verticals (from loan collection to customer service), companies are rapidly replatforming, and entire industries may flip from seat-based to outcome-based pricing within quarters—Metronome is positioned as the "payment processor" for this AI economy. MENTIONED IN THE CONVERSATION Topics: Metronome, Dropbox, Datadog, OpenAI, AI, AGI, machine learning, pricing models, financial services, business optimization, operational frameworks, analytics, financial modeling ABOUT THE FINTECH BLUEPRINT 

Bernecker Opinion
Aktien-Schnelltest inklusive Datadog, UIPath, Energiekontor & PVA TePla

Bernecker Opinion

Play Episode Listen Later Mar 25, 2025 18:45


Michael Hüsgen im Gespräch mit Oliver Kantimm ("Der Aktionärsbrief"). Hier im Beitrag gibt es die Podcast-Variante zur eigentlichen Hauptsendung im Rahmen von BerneckerTV (Aufzeichnung am 20.03.2025). Schlaglichter:Datadog - Was tun nach dem 40 % Kursrutsch?UIPath - Schwarze Zahlen weit entfernt?Energiekontor - Profiteur des Infrastrukturpakets?Werblicher Hinweis auf den AktionärsbriefPVA TePlA - Ausblick als Haar in der SuppeWir wünschen gewinnbringende Impulse mit diesem Beitrag.=======Hier gibt es weitere Infos zu "Der Aktionärsbrief":https://www.bernecker.info/aktionaersbrief=======Lust auf noch mehr Sendungen im Bernecker.TV? Noch mehr unterschiedliche Experten? Infos zu Bernecker.TV:https://www.bernecker.info/bernecker-tv=======Anmeldung zum kostenlosen Experten-Newsletter der Bernecker-Redaktion über unsere Website:https://www.bernecker.info/newsletter

The MongoDB Podcast
EP. 257 Optimizing MongoDB: Deep Dive into Database Performance, Reliability, and Cost Efficiency with Observability Tools

The MongoDB Podcast

Play Episode Listen Later Feb 28, 2025 66:07


In this episode of MongoDB TV, join Shane McAllister along with MongoDB experts Sabina Friden and Frank Sun as they explore the powerful observability suite within MongoDB Atlas. Discover how these tools can help you optimize database performance, reduce costs, and ensure reliability for your applications. From customizable alerts and query insights to performance advisors and seamless integrations with enterprise tools like Datadog and Prometheus, this episode covers it all. Whether you're a developer, database administrator, or just getting started with MongoDB, learn how to leverage these observability tools to gain deep insights into your database operations and improve your application's efficiency. Tune in for a live demo showcasing how MongoDB's observability suite can transform your database management experience. Perfect for anyone looking to enhance their MongoDB skills and take their database performance to the next level.

FedScoop Radio
Achieving zero trust with full-scale observability | Datadog's Emilio Escobar

FedScoop Radio

Play Episode Listen Later Feb 18, 2025 18:12


Datadog CISO Emilio Escobar joins SNG host Wyatt Kash in a sponsored podcast discussion on how federal agencies must leverage comprehensive observability and monitoring to overcome escalating cyber threats. This segment was sponsored by Datadog.

Doppelgänger Tech Talk
OpenAI & Anthropic Roadmaps | Earnings von AppLovin Reddit TheTradeDesk Airbnb Adyen #432

Doppelgänger Tech Talk

Play Episode Listen Later Feb 14, 2025 71:17


Während Apple hat nächste Woche ein neues Familienmitglied vorstellt, lässt sich OpenAI Zeit. Anthropic hat sich angeschaut wie wir AI nutzen und prognostiziert $34,5 Milliarden Umsatz in 2027. Arm baut Chips für Meta. Für Chris schauen wir uns endlich die Zahlen von AppLovin an. Dazu gibt es noch Earnings von Reddit, TheTradeDesk, Airbnb, Adyen. Entdecke die Angebote unserer Werbepartner auf doppelgaenger.io/werbung. Vielen Dank! Philipp Glöckler und Philipp Klöckner sprechen heute über: (00:00:00) Apple (00:02:50) OpenAI (00:05:15) Anthropic (00:17:30) Arm (00:19:30) AppLovin (00:27:45) Reddit (00:37:20) Adyen (00:40:00) TheTradeDesk (00:43:10) Robinhood (00:50:00) Coinbase (00:50:40) Airbnb (00:58:50) Datadog (00:59:00) Boulevard Corner Shownotes OpenAI verschiebt sein o3-KI-Modell zugunsten einer „einheitlichen“ Version für die nächste Generation TechCrunch Anthropic prognostiziert rasantes Wachstum auf 34,5 Milliarden Dollar Umsatz im Jahr 2027 The Information Der „Index“ von Anthropic verfolgt die KI-Wirtschaft Axios Arm sichert sich Meta als ersten Kunden für ehrgeiziges neues Chip-Projekt Reuters Google Maps zeigt tatsächlich „(Golf von Amerika)“ an, während Google ... lilyraynyc

Les Matinales de KPMG
Quand le monde de l'éducation se mobilise pour former les talents à l'intelligence artificielle

Les Matinales de KPMG

Play Episode Listen Later Feb 14, 2025 20:45


 A un moment où les annonces spectaculaires autour de l'intelligence artificielle se succèdent, les enjeux pour les entreprises ne sont plus uniquement d'ordre financier : ils sont également humains. Plus que jamais, disposer des compétences adéquates dans ce domaine est stratégique pour les DRH. Dans cette course à l'innovation, écoles et universités s'adaptent et imaginent de nouveaux cursus. Au programme de cette édition :Quels sont les principaux enjeux de la formation liée à l'intelligence artificielle ? Francesca Bugiotti et Guillaume Lecué, Directeurs académiques du Bachelor AIDAMS ESSEC & CentraleSupélec, et Capucine Marteau, Product manager chez Datadog répondent à cette question-clé à notre microNotre coup de cœur culturel nous emmène au Théâtre de la Renaissance…

Ransquawk Rundown, Daily Podcast
Europe Market Open: Geopolitics in the driving seat ahead of US data

Ransquawk Rundown, Daily Podcast

Play Episode Listen Later Feb 13, 2025 2:35


APAC stocks traded somewhat mixed albeit with a mostly positive bias among the major indices following the two-way price action across global markets owing to hot US CPI data and geopolitical optimism.US President Trump posted on Truth that he had a lengthy and highly productive phone call with Russian President Putin, and they agreed to have their respective teams start negotiations immediately. Trump then said he spoke to Ukrainian President Zelensky and the conversation went very well.US President Trump did not sign reciprocal tariffs order on Wednesday after stating that he may, while the White House schedule showed President Trump is to sign executive orders on Thursday at 13:00EST/18:00GMT.Fed Chair Powell offered a note of caution on the latest CPI reading and said the Fed targets PCE inflation, which is a better measure, and stated they will know what PCE readings are late on Thursday after the PPI data.European equity futures indicate a higher cash market open with Euro Stoxx 50 futures up by 1.1% after the cash market closed with gains of 0.3% on Wednesday.Looking ahead, highlights include German Final CPI, UK GDP Estimate and Services, Swiss CPI, US Jobless Claims, PPI, IEA OMR, Supply from Italy & US, Comments from ECB's Cipollone.Earnings from Datadog, Baxter, Deere, Duke Energy, GE Healthcare, PG&E, Coinbase, Draftkings, Applied Materials, Airbnb, Palo Alto, Roku, Wynn, Siemens, Delivery Hero, Commerzbank, Nestle, Orange, British American Tobacco, Unilever, Barclays & Moncler.Read the full report covering Equities, Forex, Fixed Income, Commodites and more on Newsquawk

Ransquawk Rundown, Daily Podcast
US Market Open: Crude subdued with continued focus on geopolitics, USD lower into PPI & Executive Orders

Ransquawk Rundown, Daily Podcast

Play Episode Listen Later Feb 13, 2025 3:17


US President Trump did not sign reciprocal tariffs order on Wednesday after stating that he may, while the White House schedule showed President Trump is to sign executive orders on Thursday at 13:00EST/18:00GMT.Stocks mostly firmer on constructive geopolitical updates; US futures are mixed ahead of PPI.USD softer as markets weigh potential Ukraine peace deal and lack of reciprocal tariffs (so far).Bonds attempt to recoup CPI-driven losses into PPI though geopols is capping.Crude continues the Russia/Ukraine downside seen in the prior session; reports suggested Israel/Hamas had come to an understanding, but this was subsequently denied by Israeli PM Netanyahu's Office.Looking ahead, US Jobless Claims, PPI, Supply from the US. Earnings from Datadog, Baxter, Deere, Duke Energy, GE Healthcare, PG&E, Coinbase, Draftkings, Applied Materials, Airbnb, Palo Alto, Roku, Wynn.Read the full report covering Equities, Forex, Fixed Income, Commodites and more on Newsquawk

CFO Thought Leader
1070: From Finance Leader to Entreprenuer: A CFO's Journey to the CEO Office | Damon Fletcher, CEO, Caliper

CFO Thought Leader

Play Episode Listen Later Feb 9, 2025 42:24


Like many seasoned finance executives, Damon Fletcher saw Snowflake as a game-changer in cloud-based data management. While a senior finance executive at Tableau, he championed its adoption, recognizing its ability to scale analytics and streamline enterprise data operations. But he also discovered a challenge familiar to many finance leaders—the hidden costs that come with cloud consumption-based pricing.At Tableau, Fletcher tells us, the company's Snowflake costs grew exponentially, mirroring a broader trend in tech where companies struggle to control cloud spend. This realization led Fletcher beyond the CFO office. In 2023, he co-founded Caliper, a company dedicated to bringing greater cost transparency and AI-powered efficiency to cloud spending.Fletcher tells us that AI is central to Caliper's approach. The platform leverages machine learning forecasting to predict cloud usage trends and generative AI to surface actionable cost-saving recommendations. Unlike traditional cloud cost tools, Caliper provides deep insights across Snowflake, AWS, and Datadog, allowing finance and DevOps teams to pinpoint inefficiencies in real time.

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

Did you know that adding a simple Code Interpreter took o3 from 9.2% to 32% on FrontierMath? The Latent Space crew is hosting a hack night Feb 11th in San Francisco focused on CodeGen use cases, co-hosted with E2B and Edge AGI; watch E2B's new workshop and RSVP here!We're happy to announce that today's guest Samuel Colvin will be teaching his very first Pydantic AI workshop at the newly announced AI Engineer NYC Workshops day on Feb 22! 25 tickets left.If you're a Python developer, it's very likely that you've heard of Pydantic. Every month, it's downloaded >300,000,000 times, making it one of the top 25 PyPi packages. OpenAI uses it in its SDK for structured outputs, it's at the core of FastAPI, and if you've followed our AI Engineer Summit conference, Jason Liu of Instructor has given two great talks about it: “Pydantic is all you need” and “Pydantic is STILL all you need”. Now, Samuel Colvin has raised $17M from Sequoia to turn Pydantic from an open source project to a full stack AI engineer platform with Logfire, their observability platform, and PydanticAI, their new agent framework.Logfire: bringing OTEL to AIOpenTelemetry recently merged Semantic Conventions for LLM workloads which provides standard definitions to track performance like gen_ai.server.time_per_output_token. In Sam's view at least 80% of new apps being built today have some sort of LLM usage in them, and just like web observability platform got replaced by cloud-first ones in the 2010s, Logfire wants to do the same for AI-first apps. If you're interested in the technical details, Logfire migrated away from Clickhouse to Datafusion for their backend. We spent some time on the importance of picking open source tools you understand and that you can actually contribute to upstream, rather than the more popular ones; listen in ~43:19 for that part.Agents are the killer app for graphsPydantic AI is their attempt at taking a lot of the learnings that LangChain and the other early LLM frameworks had, and putting Python best practices into it. At an API level, it's very similar to the other libraries: you can call LLMs, create agents, do function calling, do evals, etc.They define an “Agent” as a container with a system prompt, tools, structured result, and an LLM. Under the hood, each Agent is now a graph of function calls that can orchestrate multi-step LLM interactions. You can start simple, then move toward fully dynamic graph-based control flow if needed.“We were compelled enough by graphs once we got them right that our agent implementation [...] is now actually a graph under the hood.”Why Graphs?* More natural for complex or multi-step AI workflows.* Easy to visualize and debug with mermaid diagrams.* Potential for distributed runs, or “waiting days” between steps in certain flows.In parallel, you see folks like Emil Eifrem of Neo4j talk about GraphRAG as another place where graphs fit really well in the AI stack, so it might be time for more people to take them seriously.Full Video EpisodeLike and subscribe!Chapters* 00:00:00 Introductions* 00:00:24 Origins of Pydantic* 00:05:28 Pydantic's AI moment * 00:08:05 Why build a new agents framework?* 00:10:17 Overview of Pydantic AI* 00:12:33 Becoming a believer in graphs* 00:24:02 God Model vs Compound AI Systems* 00:28:13 Why not build an LLM gateway?* 00:31:39 Programmatic testing vs live evals* 00:35:51 Using OpenTelemetry for AI traces* 00:43:19 Why they don't use Clickhouse* 00:48:34 Competing in the observability space* 00:50:41 Licensing decisions for Pydantic and LogFire* 00:51:48 Building Pydantic.run* 00:55:24 Marimo and the future of Jupyter notebooks* 00:57:44 London's AI sceneShow Notes* Sam Colvin* Pydantic* Pydantic AI* Logfire* Pydantic.run* Zod* E2B* Arize* Langsmith* Marimo* Prefect* GLA (Google Generative Language API)* OpenTelemetry* Jason Liu* Sebastian Ramirez* Bogomil Balkansky* Hood Chatham* Jeremy Howard* Andrew LambTranscriptAlessio [00:00:03]: Hey, everyone. Welcome to the Latent Space podcast. This is Alessio, partner and CTO at Decibel Partners, and I'm joined by my co-host Swyx, founder of Smol AI.Swyx [00:00:12]: Good morning. And today we're very excited to have Sam Colvin join us from Pydantic AI. Welcome. Sam, I heard that Pydantic is all we need. Is that true?Samuel [00:00:24]: I would say you might need Pydantic AI and Logfire as well, but it gets you a long way, that's for sure.Swyx [00:00:29]: Pydantic almost basically needs no introduction. It's almost 300 million downloads in December. And obviously, in the previous podcasts and discussions we've had with Jason Liu, he's been a big fan and promoter of Pydantic and AI.Samuel [00:00:45]: Yeah, it's weird because obviously I didn't create Pydantic originally for uses in AI, it predates LLMs. But it's like we've been lucky that it's been picked up by that community and used so widely.Swyx [00:00:58]: Actually, maybe we'll hear it. Right from you, what is Pydantic and maybe a little bit of the origin story?Samuel [00:01:04]: The best name for it, which is not quite right, is a validation library. And we get some tension around that name because it doesn't just do validation, it will do coercion by default. We now have strict mode, so you can disable that coercion. But by default, if you say you want an integer field and you get in a string of 1, 2, 3, it will convert it to 123 and a bunch of other sensible conversions. And as you can imagine, the semantics around it. Exactly when you convert and when you don't, it's complicated, but because of that, it's more than just validation. Back in 2017, when I first started it, the different thing it was doing was using type hints to define your schema. That was controversial at the time. It was genuinely disapproved of by some people. I think the success of Pydantic and libraries like FastAPI that build on top of it means that today that's no longer controversial in Python. And indeed, lots of other people have copied that route, but yeah, it's a data validation library. It uses type hints for the for the most part and obviously does all the other stuff you want, like serialization on top of that. But yeah, that's the core.Alessio [00:02:06]: Do you have any fun stories on how JSON schemas ended up being kind of like the structure output standard for LLMs? And were you involved in any of these discussions? Because I know OpenAI was, you know, one of the early adopters. So did they reach out to you? Was there kind of like a structure output console in open source that people were talking about or was it just a random?Samuel [00:02:26]: No, very much not. So I originally. Didn't implement JSON schema inside Pydantic and then Sebastian, Sebastian Ramirez, FastAPI came along and like the first I ever heard of him was over a weekend. I got like 50 emails from him or 50 like emails as he was committing to Pydantic, adding JSON schema long pre version one. So the reason it was added was for OpenAPI, which is obviously closely akin to JSON schema. And then, yeah, I don't know why it was JSON that got picked up and used by OpenAI. It was obviously very convenient for us. That's because it meant that not only can you do the validation, but because Pydantic will generate you the JSON schema, it will it kind of can be one source of source of truth for structured outputs and tools.Swyx [00:03:09]: Before we dive in further on the on the AI side of things, something I'm mildly curious about, obviously, there's Zod in JavaScript land. Every now and then there is a new sort of in vogue validation library that that takes over for quite a few years and then maybe like some something else comes along. Is Pydantic? Is it done like the core Pydantic?Samuel [00:03:30]: I've just come off a call where we were redesigning some of the internal bits. There will be a v3 at some point, which will not break people's code half as much as v2 as in v2 was the was the massive rewrite into Rust, but also fixing all the stuff that was broken back from like version zero point something that we didn't fix in v1 because it was a side project. We have plans to move some of the basically store the data in Rust types after validation. Not completely. So we're still working to design the Pythonic version of it, in order for it to be able to convert into Python types. So then if you were doing like validation and then serialization, you would never have to go via a Python type we reckon that can give us somewhere between three and five times another three to five times speed up. That's probably the biggest thing. Also, like changing how easy it is to basically extend Pydantic and define how particular types, like for example, NumPy arrays are validated and serialized. But there's also stuff going on. And for example, Jitter, the JSON library in Rust that does the JSON parsing, has SIMD implementation at the moment only for AMD64. So we can add that. We need to go and add SIMD for other instruction sets. So there's a bunch more we can do on performance. I don't think we're going to go and revolutionize Pydantic, but it's going to continue to get faster, continue, hopefully, to allow people to do more advanced things. We might add a binary format like CBOR for serialization for when you'll just want to put the data into a database and probably load it again from Pydantic. So there are some things that will come along, but for the most part, it should just get faster and cleaner.Alessio [00:05:04]: From a focus perspective, I guess, as a founder too, how did you think about the AI interest rising? And then how do you kind of prioritize, okay, this is worth going into more, and we'll talk about Pydantic AI and all of that. What was maybe your early experience with LLAMP, and when did you figure out, okay, this is something we should take seriously and focus more resources on it?Samuel [00:05:28]: I'll answer that, but I'll answer what I think is a kind of parallel question, which is Pydantic's weird, because Pydantic existed, obviously, before I was starting a company. I was working on it in my spare time, and then beginning of 22, I started working on the rewrite in Rust. And I worked on it full-time for a year and a half, and then once we started the company, people came and joined. And it was a weird project, because that would never go away. You can't get signed off inside a startup. Like, we're going to go off and three engineers are going to work full-on for a year in Python and Rust, writing like 30,000 lines of Rust just to release open-source-free Python library. The result of that has been excellent for us as a company, right? As in, it's made us remain entirely relevant. And it's like, Pydantic is not just used in the SDKs of all of the AI libraries, but I can't say which one, but one of the big foundational model companies, when they upgraded from Pydantic v1 to v2, their number one internal model... The metric of performance is time to first token. That went down by 20%. So you think about all of the actual AI going on inside, and yet at least 20% of the CPU, or at least the latency inside requests was actually Pydantic, which shows like how widely it's used. So we've benefited from doing that work, although it didn't, it would have never have made financial sense in most companies. In answer to your question about like, how do we prioritize AI, I mean, the honest truth is we've spent a lot of the last year and a half building. Good general purpose observability inside LogFire and making Pydantic good for general purpose use cases. And the AI has kind of come to us. Like we just, not that we want to get away from it, but like the appetite, uh, both in Pydantic and in LogFire to go and build with AI is enormous because it kind of makes sense, right? Like if you're starting a new greenfield project in Python today, what's the chance that you're using GenAI 80%, let's say, globally, obviously it's like a hundred percent in California, but even worldwide, it's probably 80%. Yeah. And so everyone needs that stuff. And there's so much yet to be figured out so much like space to do things better in the ecosystem in a way that like to go and implement a database that's better than Postgres is a like Sisyphean task. Whereas building, uh, tools that are better for GenAI than some of the stuff that's about now is not very difficult. Putting the actual models themselves to one side.Alessio [00:07:40]: And then at the same time, then you released Pydantic AI recently, which is, uh, um, you know, agent framework and early on, I would say everybody like, you know, Langchain and like, uh, Pydantic kind of like a first class support, a lot of these frameworks, we're trying to use you to be better. What was the decision behind we should do our own framework? Were there any design decisions that you disagree with any workloads that you think people didn't support? Well,Samuel [00:08:05]: it wasn't so much like design and workflow, although I think there were some, some things we've done differently. Yeah. I think looking in general at the ecosystem of agent frameworks, the engineering quality is far below that of the rest of the Python ecosystem. There's a bunch of stuff that we have learned how to do over the last 20 years of building Python libraries and writing Python code that seems to be abandoned by people when they build agent frameworks. Now I can kind of respect that, particularly in the very first agent frameworks, like Langchain, where they were literally figuring out how to go and do this stuff. It's completely understandable that you would like basically skip some stuff.Samuel [00:08:42]: I'm shocked by the like quality of some of the agent frameworks that have come out recently from like well-respected names, which it just seems to be opportunism and I have little time for that, but like the early ones, like I think they were just figuring out how to do stuff and just as lots of people have learned from Pydantic, we were able to learn a bit from them. I think from like the gap we saw and the thing we were frustrated by was the production readiness. And that means things like type checking, even if type checking makes it hard. Like Pydantic AI, I will put my hand up now and say it has a lot of generics and you need to, it's probably easier to use it if you've written a bit of Rust and you really understand generics, but like, and that is, we're not claiming that that makes it the easiest thing to use in all cases, we think it makes it good for production applications in big systems where type checking is a no-brainer in Python. But there are also a bunch of stuff we've learned from maintaining Pydantic over the years that we've gone and done. So every single example in Pydantic AI's documentation is run on Python. As part of tests and every single print output within an example is checked during tests. So it will always be up to date. And then a bunch of things that, like I say, are standard best practice within the rest of the Python ecosystem, but I'm not followed surprisingly by some AI libraries like coverage, linting, type checking, et cetera, et cetera, where I think these are no-brainers, but like weirdly they're not followed by some of the other libraries.Alessio [00:10:04]: And can you just give an overview of the framework itself? I think there's kind of like the. LLM calling frameworks, there are the multi-agent frameworks, there's the workflow frameworks, like what does Pydantic AI do?Samuel [00:10:17]: I glaze over a bit when I hear all of the different sorts of frameworks, but I like, and I will tell you when I built Pydantic, when I built Logfire and when I built Pydantic AI, my methodology is not to go and like research and review all of the other things. I kind of work out what I want and I go and build it and then feedback comes and we adjust. So the fundamental building block of Pydantic AI is agents. The exact definition of agents and how you want to define them. is obviously ambiguous and our things are probably sort of agent-lit, not that we would want to go and rename them to agent-lit, but like the point is you probably build them together to build something and most people will call an agent. So an agent in our case has, you know, things like a prompt, like system prompt and some tools and a structured return type if you want it, that covers the vast majority of cases. There are situations where you want to go further and the most complex workflows where you want graphs and I resisted graphs for quite a while. I was sort of of the opinion you didn't need them and you could use standard like Python flow control to do all of that stuff. I had a few arguments with people, but I basically came around to, yeah, I can totally see why graphs are useful. But then we have the problem that by default, they're not type safe because if you have a like add edge method where you give the names of two different edges, there's no type checking, right? Even if you go and do some, I'm not, not all the graph libraries are AI specific. So there's a, there's a graph library called, but it allows, it does like a basic runtime type checking. Ironically using Pydantic to try and make up for the fact that like fundamentally that graphs are not typed type safe. Well, I like Pydantic, but it did, that's not a real solution to have to go and run the code to see if it's safe. There's a reason that starting type checking is so powerful. And so we kind of, from a lot of iteration eventually came up with a system of using normally data classes to define nodes where you return the next node you want to call and where we're able to go and introspect the return type of a node to basically build the graph. And so the graph is. Yeah. Inherently type safe. And once we got that right, I, I wasn't, I'm incredibly excited about graphs. I think there's like masses of use cases for them, both in gen AI and other development, but also software's all going to have interact with gen AI, right? It's going to be like web. There's no longer be like a web department in a company is that there's just like all the developers are building for web building with databases. The same is going to be true for gen AI.Alessio [00:12:33]: Yeah. I see on your docs, you call an agent, a container that contains a system prompt function. Tools, structure, result, dependency type model, and then model settings. Are the graphs in your mind, different agents? Are they different prompts for the same agent? What are like the structures in your mind?Samuel [00:12:52]: So we were compelled enough by graphs once we got them right, that we actually merged the PR this morning. That means our agent implementation without changing its API at all is now actually a graph under the hood as it is built using our graph library. So graphs are basically a lower level tool that allow you to build these complex workflows. Our agents are technically one of the many graphs you could go and build. And we just happened to build that one for you because it's a very common, commonplace one. But obviously there are cases where you need more complex workflows where the current agent assumptions don't work. And that's where you can then go and use graphs to build more complex things.Swyx [00:13:29]: You said you were cynical about graphs. What changed your mind specifically?Samuel [00:13:33]: I guess people kept giving me examples of things that they wanted to use graphs for. And my like, yeah, but you could do that in standard flow control in Python became a like less and less compelling argument to me because I've maintained those systems that end up with like spaghetti code. And I could see the appeal of this like structured way of defining the workflow of my code. And it's really neat that like just from your code, just from your type hints, you can get out a mermaid diagram that defines exactly what can go and happen.Swyx [00:14:00]: Right. Yeah. You do have very neat implementation of sort of inferring the graph from type hints, I guess. Yeah. Is what I would call it. Yeah. I think the question always is I have gone back and forth. I used to work at Temporal where we would actually spend a lot of time complaining about graph based workflow solutions like AWS step functions. And we would actually say that we were better because you could use normal control flow that you already knew and worked with. Yours, I guess, is like a little bit of a nice compromise. Like it looks like normal Pythonic code. But you just have to keep in mind what the type hints actually mean. And that's what we do with the quote unquote magic that the graph construction does.Samuel [00:14:42]: Yeah, exactly. And if you look at the internal logic of actually running a graph, it's incredibly simple. It's basically call a node, get a node back, call that node, get a node back, call that node. If you get an end, you're done. We will add in soon support for, well, basically storage so that you can store the state between each node that's run. And then the idea is you can then distribute the graph and run it across computers. And also, I mean, the other weird, the other bit that's really valuable is across time. Because it's all very well if you look at like lots of the graph examples that like Claude will give you. If it gives you an example, it gives you this lovely enormous mermaid chart of like the workflow, for example, managing returns if you're an e-commerce company. But what you realize is some of those lines are literally one function calls another function. And some of those lines are wait six days for the customer to print their like piece of paper and put it in the post. And if you're writing like your demo. Project or your like proof of concept, that's fine because you can just say, and now we call this function. But when you're building when you're in real in real life, that doesn't work. And now how do we manage that concept to basically be able to start somewhere else in the in our code? Well, this graph implementation makes it incredibly easy because you just pass the node that is the start point for carrying on the graph and it continues to run. So it's things like that where I was like, yeah, I can just imagine how things I've done in the past would be fundamentally easier to understand if we had done them with graphs.Swyx [00:16:07]: You say imagine, but like right now, this pedantic AI actually resume, you know, six days later, like you said, or is this just like a theoretical thing we can go someday?Samuel [00:16:16]: I think it's basically Q&A. So there's an AI that's asking the user a question and effectively you then call the CLI again to continue the conversation. And it basically instantiates the node and calls the graph with that node again. Now, we don't have the logic yet for effectively storing state in the database between individual nodes that we're going to add soon. But like the rest of it is basically there.Swyx [00:16:37]: It does make me think that not only are you competing with Langchain now and obviously Instructor, and now you're going into sort of the more like orchestrated things like Airflow, Prefect, Daxter, those guys.Samuel [00:16:52]: Yeah, I mean, we're good friends with the Prefect guys and Temporal have the same investors as us. And I'm sure that my investor Bogomol would not be too happy if I was like, oh, yeah, by the way, as well as trying to take on Datadog. We're also going off and trying to take on Temporal and everyone else doing that. Obviously, we're not doing all of the infrastructure of deploying that right yet, at least. We're, you know, we're just building a Python library. And like what's crazy about our graph implementation is, sure, there's a bit of magic in like introspecting the return type, you know, extracting things from unions, stuff like that. But like the actual calls, as I say, is literally call a function and get back a thing and call that. It's like incredibly simple and therefore easy to maintain. The question is, how useful is it? Well, I don't know yet. I think we have to go and find out. We have a whole. We've had a slew of people joining our Slack over the last few days and saying, tell me how good Pydantic AI is. How good is Pydantic AI versus Langchain? And I refuse to answer. That's your job to go and find that out. Not mine. We built a thing. I'm compelled by it, but I'm obviously biased. The ecosystem will work out what the useful tools are.Swyx [00:17:52]: Bogomol was my board member when I was at Temporal. And I think I think just generally also having been a workflow engine investor and participant in this space, it's a big space. Like everyone needs different functions. I think the one thing that I would say like yours, you know, as a library, you don't have that much control of it over the infrastructure. I do like the idea that each new agents or whatever or unit of work, whatever you call that should spin up in this sort of isolated boundaries. Whereas yours, I think around everything runs in the same process. But you ideally want to sort of spin out its own little container of things.Samuel [00:18:30]: I agree with you a hundred percent. And we will. It would work now. Right. As in theory, you're just like as long as you can serialize the calls to the next node, you just have to all of the different containers basically have to have the same the same code. I mean, I'm super excited about Cloudflare workers running Python and being able to install dependencies. And if Cloudflare could only give me my invitation to the private beta of that, we would be exploring that right now because I'm super excited about that as a like compute level for some of this stuff where exactly what you're saying, basically. You can run everything as an individual. Like worker function and distribute it. And it's resilient to failure, et cetera, et cetera.Swyx [00:19:08]: And it spins up like a thousand instances simultaneously. You know, you want it to be sort of truly serverless at once. Actually, I know we have some Cloudflare friends who are listening, so hopefully they'll get in front of the line. Especially.Samuel [00:19:19]: I was in Cloudflare's office last week shouting at them about other things that frustrate me. I have a love-hate relationship with Cloudflare. Their tech is awesome. But because I use it the whole time, I then get frustrated. So, yeah, I'm sure I will. I will. I will get there soon.Swyx [00:19:32]: There's a side tangent on Cloudflare. Is Python supported at full? I actually wasn't fully aware of what the status of that thing is.Samuel [00:19:39]: Yeah. So Pyodide, which is Python running inside the browser in scripting, is supported now by Cloudflare. They basically, they're having some struggles working out how to manage, ironically, dependencies that have binaries, in particular, Pydantic. Because these workers where you can have thousands of them on a given metal machine, you don't want to have a difference. You basically want to be able to have a share. Shared memory for all the different Pydantic installations, effectively. That's the thing they work out. They're working out. But Hood, who's my friend, who is the primary maintainer of Pyodide, works for Cloudflare. And that's basically what he's doing, is working out how to get Python running on Cloudflare's network.Swyx [00:20:19]: I mean, the nice thing is that your binary is really written in Rust, right? Yeah. Which also compiles the WebAssembly. Yeah. So maybe there's a way that you'd build... You have just a different build of Pydantic and that ships with whatever your distro for Cloudflare workers is.Samuel [00:20:36]: Yes, that's exactly what... So Pyodide has builds for Pydantic Core and for things like NumPy and basically all of the popular binary libraries. Yeah. It's just basic. And you're doing exactly that, right? You're using Rust to compile the WebAssembly and then you're calling that shared library from Python. And it's unbelievably complicated, but it works. Okay.Swyx [00:20:57]: Staying on graphs a little bit more, and then I wanted to go to some of the other features that you have in Pydantic AI. I see in your docs, there are sort of four levels of agents. There's single agents, there's agent delegation, programmatic agent handoff. That seems to be what OpenAI swarms would be like. And then the last one, graph-based control flow. Would you say that those are sort of the mental hierarchy of how these things go?Samuel [00:21:21]: Yeah, roughly. Okay.Swyx [00:21:22]: You had some expression around OpenAI swarms. Well.Samuel [00:21:25]: And indeed, OpenAI have got in touch with me and basically, maybe I'm not supposed to say this, but basically said that Pydantic AI looks like what swarms would become if it was production ready. So, yeah. I mean, like, yeah, which makes sense. Awesome. Yeah. I mean, in fact, it was specifically saying, how can we give people the same feeling that they were getting from swarms that led us to go and implement graphs? Because my, like, just call the next agent with Python code was not a satisfactory answer to people. So it was like, okay, we've got to go and have a better answer for that. It's not like, let us to get to graphs. Yeah.Swyx [00:21:56]: I mean, it's a minimal viable graph in some sense. What are the shapes of graphs that people should know? So the way that I would phrase this is I think Anthropic did a very good public service and also kind of surprisingly influential blog post, I would say, when they wrote Building Effective Agents. We actually have the authors coming to speak at my conference in New York, which I think you're giving a workshop at. Yeah.Samuel [00:22:24]: I'm trying to work it out. But yes, I think so.Swyx [00:22:26]: Tell me if you're not. yeah, I mean, like, that was the first, I think, authoritative view of, like, what kinds of graphs exist in agents and let's give each of them a name so that everyone is on the same page. So I'm just kind of curious if you have community names or top five patterns of graphs.Samuel [00:22:44]: I don't have top five patterns of graphs. I would love to see what people are building with them. But like, it's been it's only been a couple of weeks. And of course, there's a point is that. Because they're relatively unopinionated about what you can go and do with them. They don't suit them. Like, you can go and do lots of lots of things with them, but they don't have the structure to go and have like specific names as much as perhaps like some other systems do. I think what our agents are, which have a name and I can't remember what it is, but this basically system of like, decide what tool to call, go back to the center, decide what tool to call, go back to the center and then exit. One form of graph, which, as I say, like our agents are effectively one implementation of a graph, which is why under the hood they are now using graphs. And it'll be interesting to see over the next few years whether we end up with these like predefined graph names or graph structures or whether it's just like, yep, I built a graph or whether graphs just turn out not to match people's mental image of what they want and die away. We'll see.Swyx [00:23:38]: I think there is always appeal. Every developer eventually gets graph religion and goes, oh, yeah, everything's a graph. And then they probably over rotate and go go too far into graphs. And then they have to learn a whole bunch of DSLs. And then they're like, actually, I didn't need that. I need this. And they scale back a little bit.Samuel [00:23:55]: I'm at the beginning of that process. I'm currently a graph maximalist, although I haven't actually put any into production yet. But yeah.Swyx [00:24:02]: This has a lot of philosophical connections with other work coming out of UC Berkeley on compounding AI systems. I don't know if you know of or care. This is the Gartner world of things where they need some kind of industry terminology to sell it to enterprises. I don't know if you know about any of that.Samuel [00:24:24]: I haven't. I probably should. I should probably do it because I should probably get better at selling to enterprises. But no, no, I don't. Not right now.Swyx [00:24:29]: This is really the argument is that instead of putting everything in one model, you have more control and more maybe observability to if you break everything out into composing little models and changing them together. And obviously, then you need an orchestration framework to do that. Yeah.Samuel [00:24:47]: And it makes complete sense. And one of the things we've seen with agents is they work well when they work well. But when they. Even if you have the observability through log five that you can see what was going on, if you don't have a nice hook point to say, hang on, this is all gone wrong. You have a relatively blunt instrument of basically erroring when you exceed some kind of limit. But like what you need to be able to do is effectively iterate through these runs so that you can have your own control flow where you're like, OK, we've gone too far. And that's where one of the neat things about our graph implementation is you can basically call next in a loop rather than just running the full graph. And therefore, you have this opportunity to to break out of it. But yeah, basically, it's the same point, which is like if you have two bigger unit of work to some extent, whether or not it involves gen AI. But obviously, it's particularly problematic in gen AI. You only find out afterwards when you've spent quite a lot of time and or money when it's gone off and done done the wrong thing.Swyx [00:25:39]: Oh, drop on this. We're not going to resolve this here, but I'll drop this and then we can move on to the next thing. This is the common way that we we developers talk about this. And then the machine learning researchers look at us. And laugh and say, that's cute. And then they just train a bigger model and they wipe us out in the next training run. So I think there's a certain amount of we are fighting the bitter lesson here. We're fighting AGI. And, you know, when AGI arrives, this will all go away. Obviously, on Latent Space, we don't really discuss that because I think AGI is kind of this hand wavy concept that isn't super relevant. But I think we have to respect that. For example, you could do a chain of thoughts with graphs and you could manually orchestrate a nice little graph that does like. Reflect, think about if you need more, more inference time, compute, you know, that's the hot term now. And then think again and, you know, scale that up. Or you could train Strawberry and DeepSeq R1. Right.Samuel [00:26:32]: I saw someone saying recently, oh, they were really optimistic about agents because models are getting faster exponentially. And I like took a certain amount of self-control not to describe that it wasn't exponential. But my main point was. If models are getting faster as quickly as you say they are, then we don't need agents and we don't really need any of these abstraction layers. We can just give our model and, you know, access to the Internet, cross our fingers and hope for the best. Agents, agent frameworks, graphs, all of this stuff is basically making up for the fact that right now the models are not that clever. In the same way that if you're running a customer service business and you have loads of people sitting answering telephones, the less well trained they are, the less that you trust them, the more that you need to give them a script to go through. Whereas, you know, so if you're running a bank and you have lots of customer service people who you don't trust that much, then you tell them exactly what to say. If you're doing high net worth banking, you just employ people who you think are going to be charming to other rich people and set them off to go and have coffee with people. Right. And the same is true of models. The more intelligent they are, the less we need to tell them, like structure what they go and do and constrain the routes in which they take.Swyx [00:27:42]: Yeah. Yeah. Agree with that. So I'm happy to move on. So the other parts of Pydantic AI that are worth commenting on, and this is like my last rant, I promise. So obviously, every framework needs to do its sort of model adapter layer, which is, oh, you can easily swap from OpenAI to Cloud to Grok. You also have, which I didn't know about, Google GLA, which I didn't really know about until I saw this in your docs, which is generative language API. I assume that's AI Studio? Yes.Samuel [00:28:13]: Google don't have good names for it. So Vertex is very clear. That seems to be the API that like some of the things use, although it returns 503 about 20% of the time. So... Vertex? No. Vertex, fine. But the... Oh, oh. GLA. Yeah. Yeah.Swyx [00:28:28]: I agree with that.Samuel [00:28:29]: So we have, again, another example of like, well, I think we go the extra mile in terms of engineering is we run on every commit, at least commit to main, we run tests against the live models. Not lots of tests, but like a handful of them. Oh, okay. And we had a point last week where, yeah, GLA is a little bit better. GLA1 was failing every single run. One of their tests would fail. And we, I think we might even have commented out that one at the moment. So like all of the models fail more often than you might expect, but like that one seems to be particularly likely to fail. But Vertex is the same API, but much more reliable.Swyx [00:29:01]: My rant here is that, you know, versions of this appear in Langchain and every single framework has to have its own little thing, a version of that. I would put to you, and then, you know, this is, this can be agree to disagree. This is not needed in Pydantic AI. I would much rather you adopt a layer like Lite LLM or what's the other one in JavaScript port key. And that's their job. They focus on that one thing and they, they normalize APIs for you. All new models are automatically added and you don't have to duplicate this inside of your framework. So for example, if I wanted to use deep seek, I'm out of luck because Pydantic AI doesn't have deep seek yet.Samuel [00:29:38]: Yeah, it does.Swyx [00:29:39]: Oh, it does. Okay. I'm sorry. But you know what I mean? Should this live in your code or should it live in a layer that's kind of your API gateway that's a defined piece of infrastructure that people have?Samuel [00:29:49]: And I think if a company who are well known, who are respected by everyone had come along and done this at the right time, maybe we should have done it a year and a half ago and said, we're going to be the universal AI layer. That would have been a credible thing to do. I've heard varying reports of Lite LLM is the truth. And it didn't seem to have exactly the type safety that we needed. Also, as I understand it, and again, I haven't looked into it in great detail. Part of their business model is proxying the request through their, through their own system to do the generalization. That would be an enormous put off to an awful lot of people. Honestly, the truth is I don't think it is that much work unifying the model. I get where you're coming from. I kind of see your point. I think the truth is that everyone is centralizing around open AIs. Open AI's API is the one to do. So DeepSeq support that. Grok with OK support that. Ollama also does it. I mean, if there is that library right now, it's more or less the open AI SDK. And it's very high quality. It's well type checked. It uses Pydantic. So I'm biased. But I mean, I think it's pretty well respected anyway.Swyx [00:30:57]: There's different ways to do this. Because also, it's not just about normalizing the APIs. You have to do secret management and all that stuff.Samuel [00:31:05]: Yeah. And there's also. There's Vertex and Bedrock, which to one extent or another, effectively, they host multiple models, but they don't unify the API. But they do unify the auth, as I understand it. Although we're halfway through doing Bedrock. So I don't know about it that well. But they're kind of weird hybrids because they support multiple models. But like I say, the auth is centralized.Swyx [00:31:28]: Yeah, I'm surprised they don't unify the API. That seems like something that I would do. You know, we can discuss all this all day. There's a lot of APIs. I agree.Samuel [00:31:36]: It would be nice if there was a universal one that we didn't have to go and build.Alessio [00:31:39]: And I guess the other side of, you know, routing model and picking models like evals. How do you actually figure out which one you should be using? I know you have one. First of all, you have very good support for mocking in unit tests, which is something that a lot of other frameworks don't do. So, you know, my favorite Ruby library is VCR because it just, you know, it just lets me store the HTTP requests and replay them. That part I'll kind of skip. I think you are busy like this test model. We're like just through Python. You try and figure out what the model might respond without actually calling the model. And then you have the function model where people can kind of customize outputs. Any other fun stories maybe from there? Or is it just what you see is what you get, so to speak?Samuel [00:32:18]: On those two, I think what you see is what you get. On the evals, I think watch this space. I think it's something that like, again, I was somewhat cynical about for some time. Still have my cynicism about some of the well, it's unfortunate that so many different things are called evals. It would be nice if we could agree. What they are and what they're not. But look, I think it's a really important space. I think it's something that we're going to be working on soon, both in Pydantic AI and in LogFire to try and support better because it's like it's an unsolved problem.Alessio [00:32:45]: Yeah, you do say in your doc that anyone who claims to know for sure exactly how your eval should be defined can safely be ignored.Samuel [00:32:52]: We'll delete that sentence when we tell people how to do their evals.Alessio [00:32:56]: Exactly. I was like, we need we need a snapshot of this today. And so let's talk about eval. So there's kind of like the vibe. Yeah. So you have evals, which is what you do when you're building. Right. Because you cannot really like test it that many times to get statistical significance. And then there's the production eval. So you also have LogFire, which is kind of like your observability product, which I tried before. It's very nice. What are some of the learnings you've had from building an observability tool for LEMPs? And yeah, as people think about evals, even like what are the right things to measure? What are like the right number of samples that you need to actually start making decisions?Samuel [00:33:33]: I'm not the best person to answer that is the truth. So I'm not going to come in here and tell you that I think I know the answer on the exact number. I mean, we can do some back of the envelope statistics calculations to work out that like having 30 probably gets you most of the statistical value of having 200 for, you know, by definition, 15% of the work. But the exact like how many examples do you need? For example, that's a much harder question to answer because it's, you know, it's deep within the how models operate in terms of LogFire. One of the reasons we built LogFire the way we have and we allow you to write SQL directly against your data and we're trying to build the like powerful fundamentals of observability is precisely because we know we don't know the answers. And so allowing people to go and innovate on how they're going to consume that stuff and how they're going to process it is we think that's valuable. Because even if we come along and offer you an evals framework on top of LogFire, it won't be right in all regards. And we want people to be able to go and innovate and being able to write their own SQL connected to the API. And effectively query the data like it's a database with SQL allows people to innovate on that stuff. And that's what allows us to do it as well. I mean, we do a bunch of like testing what's possible by basically writing SQL directly against LogFire as any user could. I think the other the other really interesting bit that's going on in observability is OpenTelemetry is centralizing around semantic attributes for GenAI. So it's a relatively new project. A lot of it's still being added at the moment. But basically the idea that like. They unify how both SDKs and or agent frameworks send observability data to to any OpenTelemetry endpoint. And so, again, we can go and having that unification allows us to go and like basically compare different libraries, compare different models much better. That stuff's in a very like early stage of development. One of the things we're going to be working on pretty soon is basically, I suspect, GenAI will be the first agent framework that implements those semantic attributes properly. Because, again, we control and we can say this is important for observability, whereas most of the other agent frameworks are not maintained by people who are trying to do observability. With the exception of Langchain, where they have the observability platform, but they chose not to go down the OpenTelemetry route. So they're like plowing their own furrow. And, you know, they're a lot they're even further away from standardization.Alessio [00:35:51]: Can you maybe just give a quick overview of how OTEL ties into the AI workflows? There's kind of like the question of is, you know, a trace. And a span like a LLM call. Is it the agent? It's kind of like the broader thing you're tracking. How should people think about it?Samuel [00:36:06]: Yeah, so they have a PR that I think may have now been merged from someone at IBM talking about remote agents and trying to support this concept of remote agents within GenAI. I'm not particularly compelled by that because I don't think that like that's actually by any means the common use case. But like, I suppose it's fine for it to be there. The majority of the stuff in OTEL is basically defining how you would instrument. A given call to an LLM. So basically the actual LLM call, what data you would send to your telemetry provider, how you would structure that. Apart from this slightly odd stuff on remote agents, most of the like agent level consideration is not yet implemented in is not yet decided effectively. And so there's a bit of ambiguity. Obviously, what's good about OTEL is you can in the end send whatever attributes you like. But yeah, there's quite a lot of churn in that space and exactly how we store the data. I think that one of the most interesting things, though, is that if you think about observability. Traditionally, it was sure everyone would say our observability data is very important. We must keep it safe. But actually, companies work very hard to basically not have anything that sensitive in their observability data. So if you're a doctor in a hospital and you search for a drug for an STI, the sequel might be sent to the observability provider. But none of the parameters would. It wouldn't have the patient number or their name or the drug. With GenAI, that distinction doesn't exist because it's all just messed up in the text. If you have that same patient asking an LLM how to. What drug they should take or how to stop smoking. You can't extract the PII and not send it to the observability platform. So the sensitivity of the data that's going to end up in observability platforms is going to be like basically different order of magnitude to what's in what you would normally send to Datadog. Of course, you can make a mistake and send someone's password or their card number to Datadog. But that would be seen as a as a like mistake. Whereas in GenAI, a lot of data is going to be sent. And I think that's why companies like Langsmith and are trying hard to offer observability. On prem, because there's a bunch of companies who are happy for Datadog to be cloud hosted, but want self-hosted self-hosting for this observability stuff with GenAI.Alessio [00:38:09]: And are you doing any of that today? Because I know in each of the spans you have like the number of tokens, you have the context, you're just storing everything. And then you're going to offer kind of like a self-hosting for the platform, basically. Yeah. Yeah.Samuel [00:38:23]: So we have scrubbing roughly equivalent to what the other observability platforms have. So if we, you know, if we see password as the key, we won't send the value. But like, like I said, that doesn't really work in GenAI. So we're accepting we're going to have to store a lot of data and then we'll offer self-hosting for those people who can afford it and who need it.Alessio [00:38:42]: And then this is, I think, the first time that most of the workloads performance is depending on a third party. You know, like if you're looking at Datadog data, usually it's your app that is driving the latency and like the memory usage and all of that. Here you're going to have spans that maybe take a long time to perform because the GLA API is not working or because OpenAI is kind of like overwhelmed. Do you do anything there since like the provider is almost like the same across customers? You know, like, are you trying to surface these things for people and say, hey, this was like a very slow span, but actually all customers using OpenAI right now are seeing the same thing. So maybe don't worry about it or.Samuel [00:39:20]: Not yet. We do a few things that people don't generally do in OTA. So we send. We send information at the beginning. At the beginning of a trace as well as sorry, at the beginning of a span, as well as when it finishes. By default, OTA only sends you data when the span finishes. So if you think about a request which might take like 20 seconds, even if some of the intermediate spans finished earlier, you can't basically place them on the page until you get the top level span. And so if you're using standard OTA, you can't show anything until those requests are finished. When those requests are taking a few hundred milliseconds, it doesn't really matter. But when you're doing Gen AI calls or when you're like running a batch job that might take 30 minutes. That like latency of not being able to see the span is like crippling to understanding your application. And so we've we do a bunch of slightly complex stuff to basically send data about a span as it starts, which is closely related. Yeah.Alessio [00:40:09]: Any thoughts on all the other people trying to build on top of OpenTelemetry in different languages, too? There's like the OpenLEmetry project, which doesn't really roll off the tongue. But how do you see the future of these kind of tools? Is everybody going to have to build? Why does everybody want to build? They want to build their own open source observability thing to then sell?Samuel [00:40:29]: I mean, we are not going off and trying to instrument the likes of the OpenAI SDK with the new semantic attributes, because at some point that's going to happen and it's going to live inside OTEL and we might help with it. But we're a tiny team. We don't have time to go and do all of that work. So OpenLEmetry, like interesting project. But I suspect eventually most of those semantic like that instrumentation of the big of the SDKs will live, like I say, inside the main OpenTelemetry report. I suppose. What happens to the agent frameworks? What data you basically need at the framework level to get the context is kind of unclear. I don't think we know the answer yet. But I mean, I was on the, I guess this is kind of semi-public, because I was on the call with the OpenTelemetry call last week talking about GenAI. And there was someone from Arize talking about the challenges they have trying to get OpenTelemetry data out of Langchain, where it's not like natively implemented. And obviously they're having quite a tough time. And I was realizing, hadn't really realized this before, but how lucky we are to primarily be talking about our own agent framework, where we have the control rather than trying to go and instrument other people's.Swyx [00:41:36]: Sorry, I actually didn't know about this semantic conventions thing. It looks like, yeah, it's merged into main OTel. What should people know about this? I had never heard of it before.Samuel [00:41:45]: Yeah, I think it looks like a great start. I think there's some unknowns around how you send the messages that go back and forth, which is kind of the most important part. It's the most important thing of all. And that is moved out of attributes and into OTel events. OTel events in turn are moving from being on a span to being their own top-level API where you send data. So there's a bunch of churn still going on. I'm impressed by how fast the OTel community is moving on this project. I guess they, like everyone else, get that this is important, and it's something that people are crying out to get instrumentation off. So I'm kind of pleasantly surprised at how fast they're moving, but it makes sense.Swyx [00:42:25]: I'm just kind of browsing through the specification. I can already see that this basically bakes in whatever the previous paradigm was. So now they have genai.usage.prompt tokens and genai.usage.completion tokens. And obviously now we have reasoning tokens as well. And then only one form of sampling, which is top-p. You're basically baking in or sort of reifying things that you think are important today, but it's not a super foolproof way of doing this for the future. Yeah.Samuel [00:42:54]: I mean, that's what's neat about OTel is you can always go and send another attribute and that's fine. It's just there are a bunch that are agreed on. But I would say, you know, to come back to your previous point about whether or not we should be relying on one centralized abstraction layer, this stuff is moving so fast that if you start relying on someone else's standard, you risk basically falling behind because you're relying on someone else to keep things up to date.Swyx [00:43:14]: Or you fall behind because you've got other things going on.Samuel [00:43:17]: Yeah, yeah. That's fair. That's fair.Swyx [00:43:19]: Any other observations just about building LogFire, actually? Let's just talk about this. So you announced LogFire. I was kind of only familiar with LogFire because of your Series A announcement. I actually thought you were making a separate company. I remember some amount of confusion with you when that came out. So to be clear, it's Pydantic LogFire and the company is one company that has kind of two products, an open source thing and an observability thing, correct? Yeah. I was just kind of curious, like any learnings building LogFire? So classic question is, do you use ClickHouse? Is this like the standard persistence layer? Any learnings doing that?Samuel [00:43:54]: We don't use ClickHouse. We started building our database with ClickHouse, moved off ClickHouse onto Timescale, which is a Postgres extension to do analytical databases. Wow. And then moved off Timescale onto DataFusion. And we're basically now building, it's DataFusion, but it's kind of our own database. Bogomil is not entirely happy that we went through three databases before we chose one. I'll say that. But like, we've got to the right one in the end. I think we could have realized that Timescale wasn't right. I think ClickHouse. They both taught us a lot and we're in a great place now. But like, yeah, it's been a real journey on the database in particular.Swyx [00:44:28]: Okay. So, you know, as a database nerd, I have to like double click on this, right? So ClickHouse is supposed to be the ideal backend for anything like this. And then moving from ClickHouse to Timescale is another counterintuitive move that I didn't expect because, you know, Timescale is like an extension on top of Postgres. Not super meant for like high volume logging. But like, yeah, tell us those decisions.Samuel [00:44:50]: So at the time, ClickHouse did not have good support for JSON. I was speaking to someone yesterday and said ClickHouse doesn't have good support for JSON and got roundly stepped on because apparently it does now. So they've obviously gone and built their proper JSON support. But like back when we were trying to use it, I guess a year ago or a bit more than a year ago, everything happened to be a map and maps are a pain to try and do like looking up JSON type data. And obviously all these attributes, everything you're talking about there in terms of the GenAI stuff. You can choose to make them top level columns if you want. But the simplest thing is just to put them all into a big JSON pile. And that was a problem with ClickHouse. Also, ClickHouse had some really ugly edge cases like by default, or at least until I complained about it a lot, ClickHouse thought that two nanoseconds was longer than one second because they compared intervals just by the number, not the unit. And I complained about that a lot. And then they caused it to raise an error and just say you have to have the same unit. Then I complained a bit more. And I think as I understand it now, they have some. They convert between units. But like stuff like that, when all you're looking at is when a lot of what you're doing is comparing the duration of spans was really painful. Also things like you can't subtract two date times to get an interval. You have to use the date sub function. But like the fundamental thing is because we want our end users to write SQL, the like quality of the SQL, how easy it is to write, matters way more to us than if you're building like a platform on top where your developers are going to write the SQL. And once it's written and it's working, you don't mind too much. So I think that's like one of the fundamental differences. The other problem that I have with the ClickHouse and Impact Timescale is that like the ultimate architecture, the like snowflake architecture of binary data in object store queried with some kind of cache from nearby. They both have it, but it's closed sourced and you only get it if you go and use their hosted versions. And so even if we had got through all the problems with Timescale or ClickHouse, we would end up like, you know, they would want to be taking their 80% margin. And then we would be wanting to take that would basically leave us less space for margin. Whereas data fusion. Properly open source, all of that same tooling is open source. And for us as a team of people with a lot of Rust expertise, data fusion, which is implemented in Rust, we can literally dive into it and go and change it. So, for example, I found that there were some slowdowns in data fusion's string comparison kernel for doing like string contains. And it's just Rust code. And I could go and rewrite the string comparison kernel to be faster. Or, for example, data fusion, when we started using it, didn't have JSON support. Obviously, as I've said, it's something we can do. It's something we needed. I was able to go and implement that in a weekend using our JSON parser that we built for Pydantic Core. So it's the fact that like data fusion is like for us the perfect mixture of a toolbox to build a database with, not a database. And we can go and implement stuff on top of it in a way that like if you were trying to do that in Postgres or in ClickHouse. I mean, ClickHouse would be easier because it's C++, relatively modern C++. But like as a team of people who are not C++ experts, that's much scarier than data fusion for us.Swyx [00:47:47]: Yeah, that's a beautiful rant.Alessio [00:47:49]: That's funny. Most people don't think they have agency on these projects. They're kind of like, oh, I should use this or I should use that. They're not really like, what should I pick so that I contribute the most back to it? You know, so but I think you obviously have an open source first mindset. So that makes a lot of sense.Samuel [00:48:05]: I think if we were probably better as a startup, a better startup and faster moving and just like headlong determined to get in front of customers as fast as possible, we should have just started with ClickHouse. I hope that long term we're in a better place for having worked with data fusion. We like we're quite engaged now with the data fusion community. Andrew Lam, who maintains data fusion, is an advisor to us. We're in a really good place now. But yeah, it's definitely slowed us down relative to just like building on ClickHouse and moving as fast as we can.Swyx [00:48:34]: OK, we're about to zoom out and do Pydantic run and all the other stuff. But, you know, my last question on LogFire is really, you know, at some point you run out sort of community goodwill just because like, oh, I use Pydantic. I love Pydantic. I'm going to use LogFire. OK, then you start entering the territory of the Datadogs, the Sentrys and the honeycombs. Yeah. So where are you going to really spike here? What differentiator here?Samuel [00:48:59]: I wasn't writing code in 2001, but I'm assuming that there were people talking about like web observability and then web observability stopped being a thing, not because the web stopped being a thing, but because all observability had to do web. If you were talking to people in 2010 or 2012, they would have talked about cloud observability. Now that's not a term because all observability is cloud first. The same is going to happen to gen AI. And so whether or not you're trying to compete with Datadog or with Arise and Langsmith, you've got to do first class. You've got to do general purpose observability with first class support for AI. And as far as I know, we're the only people really trying to do that. I mean, I think Datadog is starting in that direction. And to be honest, I think Datadog is a much like scarier company to compete with than the AI specific observability platforms. Because in my opinion, and I've also heard this from lots of customers, AI specific observability where you don't see everything else going on in your app is not actually that useful. Our hope is that we can build the first general purpose observability platform with first class support for AI. And that we have this open source heritage of putting developer experience first that other companies haven't done. For all I'm a fan of Datadog and what they've done. If you search Datadog logging Python. And you just try as a like a non-observability expert to get something up and running with Datadog and Python. It's not trivial, right? That's something Sentry have done amazingly well. But like there's enormous space in most of observability to do DX better.Alessio [00:50:27]: Since you mentioned Sentry, I'm curious how you thought about licensing and all of that. Obviously, your MIT license, you don't have any rolling license like Sentry has where you can only use an open source, like the one year old version of it. Was that a hard decision?Samuel [00:50:41]: So to be clear, LogFire is co-sourced. So Pydantic and Pydantic AI are MIT licensed and like properly open source. And then LogFire for now is completely closed source. And in fact, the struggles that Sentry have had with licensing and the like weird pushback the community gives when they take something that's closed source and make it source available just meant that we just avoided that whole subject matter. I think the other way to look at it is like in terms of either headcount or revenue or dollars in the bank. The amount of open source we do as a company is we've got to be open source. We're up there with the most prolific open source companies, like I say, per head. And so we didn't feel like we were morally obligated to make LogFire open source. We have Pydantic. Pydantic is a foundational library in Python. That and now Pydantic AI are our contribution to open source. And then LogFire is like openly for profit, right? As in we're not claiming otherwise. We're not sort of trying to walk a line if it's open source. But really, we want to make it hard to deploy. So you probably want to pay us. We're trying to be straight. That it's to pay for. We could change that at some point in the future, but it's not an immediate plan.Alessio [00:51:48]: All right. So the first one I saw this new I don't know if it's like a product you're building the Pydantic that run, which is a Python browser sandbox. What was the inspiration behind that? We talk a lot about code interpreter for lamps. I'm an investor in a company called E2B, which is a code sandbox as a service for remote execution. Yeah. What's the Pydantic that run story?Samuel [00:52:09]: So Pydantic that run is again completely open source. I have no interest in making it into a product. We just needed a sandbox to be able to demo LogFire in particular, but also Pydantic AI. So it doesn't have it yet, but I'm going to add basically a proxy to OpenAI and the other models so that you can run Pydantic AI in the browser. See how it works. Tweak the prompt, et cetera, et cetera. And we'll have some kind of limit per day of what you can spend on it or like what the spend is. The other thing we wanted to b

CaSE: Conversations about Software Engineering
New Hosts and Formats, Observability Costs and Training

CaSE: Conversations about Software Engineering

Play Episode Listen Later Feb 3, 2025 81:42


The CaSE Podcast returns with new hosts and a renewed focus on software architecture, reliability engineering, and data engineering. In this episode we start with discussing the cost of observability, sparked by Coinbase's leaked $65 million Datadog bill, raising questions about how much organizations should spend on monitoring. We also discuss the most important content of observability training for software architects. We close with Alex' current thoughts on home automation while renovating his house.

Les Cast Codeurs Podcast
LCC 321 - Les évènements écran large

Les Cast Codeurs Podcast

Play Episode Listen Later Jan 21, 2025 73:53


Arnaud et Emmanuel discutent des versions Java, font un résumé de l'ecosystème WebAssembly, discutent du nouveau Model Context Protocol, parlent d'observabilité avec notamment les Wide Events et de pleins d'autres choses encore. Enregistré le 17 janvier 2025 Téléchargement de l'épisode LesCastCodeurs-Episode–321.mp3 ou en vidéo sur YouTube. News Langages java trend par InfoQ https://www.infoq.com/articles/java-trends-report–2024/ Java 17 finalement depasse 11 et 8 ~30/33% Java 21 est à 1.4% commonhaus apparait GraalVM en early majority Spring AI et langchain4j en innovateurs SB 3 voit son adoption augmenter Un bon résumé sur WebAssembly, les différentes specs comme WASM GC, WASI, WIT, etc https://2ality.com/2025/01/webassembly-language-ecosystem.html WebAssembly (Wasm) est un format d'instructions binaires pour une machine virtuelle basée sur une pile, permettant la portabilité et l'efficacité du code. Wasm a évolué à partir d'asm.js, un sous-ensemble de JavaScript qui pouvait fonctionner à des vitesses proches de celles natives. WASI (WebAssembly System Interface) permet à Wasm de fonctionner en dehors des navigateurs Web, fournissant des API pour le système de fichiers, CLI, HTTP, etc. Le modèle de composant WebAssembly permet l'interopérabilité entre les langages Wasm à l'aide de WIT (Wasm Interface Type) et d'ABI canonique. Les composants Wasm se composent d'un module central et d'interfaces WIT pour les importations/exportations, facilitant l'interaction indépendante du langage. Les interfaces WIT décrivent les types et les fonctions, tandis que les mondes WIT définissent les capacités et les besoins d'un composant (importations/exportations). La gestion des packages Wasm est assurée par Warg, un protocole pour les registres de packages Wasm. Une enquête a montré que Rust est le langage Wasm le plus utilisé, suivi de Kotlin et de C++; de nombreux autres langages sont également en train d'émerger. Un algorithme de comptage a taille limitée ne mémoire a été inventé https://www.quantamagazine.org/computer-scientists-invent-an-efficient-new-way-to-count–20240516/ élimine un mot de manière aléatoire mais avec une probabilité connue quand il y a besoin de récupérer de l'espace cela se fait par round et on augmente la probabilité de suppression à chaque round donc au final, ne nombre de mots / la probabilité d'avoir été éliminé donne une mesure approximative mais plutot précise Librairies Les contributions Spring passent du CLA au DCO https://spring.io/blog/2025/01/06/hello-dco-goodbye-cla-simplifying-contributions-to-spring d'abord manuel amis meme automatisé le CLA est une document legal complexe qui peut limiter les contribuitions le DCO vient le Linux je crois et est super simple accord que la licence de la conmtrib est celle du projet accord que le code est public et distribué en perpetuité s'appuie sur les -s de git pour le sign off Ecrire un serveur MCP en Quarkus https://quarkus.io/blog/mcp-server/ MCP est un protocol proposé paor Antropic pour integrer des outils orchestrables par les LLMs MCP est frais et va plus loin que les outils offre la notion de resource (file), de functions (tools), et de proimpts pre-built pour appeler l'outil de la meilleure façon On en reparlera a pres avec les agent dans un article suivant il y a une extension Quarkus pour simplifier le codage un article plus detaillé sur l'integration Quarkus https://quarkus.io/blog/quarkus-langchain4j-mcp/ GreenMail un mini mail server en java https://greenmail-mail-test.github.io/greenmail/#features-api Utile pour les tests d'integration Supporte SMTP, POP3 et IMAP avec TLS/SSL Propose des integrations JUnit, Spring Une mini UI et des APIs REST permettent d'interagir avec le serveur si par exemple vous le partagé dans un container (il n'y a pas d'integration TestContainer existante mais elle n'est pas compliquée à écrire) Infrastructure Docker Bake in a visual way https://dev.to/aurelievache/understanding-docker-part–47-docker-bake–4p05 docker back propose d'utiliser des fichiers de configuration (format HCL) pour lancer ses builds d'images et docker compose en gros voyez ce DSL comme un Makefile très simplifié pour les commandes docker qui souvent peuvent avoir un peu trop de paramètres Datadog continue de s'etendre avec l'acquisition de Quickwit https://www.datadoghq.com/blog/datadog-acquires-quickwit/ Solution open-source de recherche des logs qui peut être déployée on-premise et dans le cloud https://quickwit.io/ Les logs ne quittent plus votre environment ce qui permet de répondre à des besoins de sécurité, privacy et réglementaire Web 33 concepts en javascript https://github.com/leonardomso/33-js-concepts Call Stack, Primitive Types, Value Types and Reference Types, Implicit, Explicit, Nominal, Structuring and Duck Typing, == vs === vs typeof, Function Scope, Block Scope and Lexical Scope, Expression vs Statement, IIFE, Modules and Namespaces, Message Queue and Event Loop, setTimeout, setInterval and requestAnimationFrame, JavaScript Engines, Bitwise Operators, Type Arrays and Array Buffers, DOM and Layout Trees, Factories and Classes, this, call, apply and bind, new, Constructor, instanceof and Instances, Prototype Inheritance and Prototype Chain, Object.create and Object.assign, map, reduce, filter, Pure Functions, Side Effects, State Mutation and Event Propagation, Closures, High Order Functions, Recursion, Collections and Generators, Promises, async/await, Data Structures, Expensive Operation and Big O Notation, Algorithms, Inheritance, Polymorphism and Code Reuse, Design Patterns, Partial Applications, Currying, Compose and Pipe, Clean Code Data et Intelligence Artificielle Phi 4 et les small language models https://techcommunity.microsoft.com/blog/aiplatformblog/introducing-phi–4-microsoft%e2%80%99s-newest-small-language-model-specializing-in-comple/4357090 Phi 4 un SML pour les usages locaux notamment 14B de parametres belle progression de ~20 points sur un score aggregé et qui le rapproche de Llama 3.3 et ses 70B de parametres bon en math (data set synthétique) Comment utiliser Gemini 2.0 Flash Thinking (le modèle de Google qui fait du raisonnement à la sauce chain of thought) en Java avec LangChain4j https://glaforge.dev/posts/2024/12/20/lets-think-with-gemini–2-thinking-mode-and-langchain4j/ Google a sorti Gemini 2.0 Flash, un petit modèle de la famille Gemini the “thinking mode” simule les cheminements de pensée (Chain of thoughts etc) décompose beaucoup plus les taches coplexes en plusiewurs taches un exemple est montré sur le modele se battant avec le probleme Les recommendations d'Antropic sur les systèmes d'agents https://www.anthropic.com/research/building-effective-agents défini les agents et les workflow Ne recommence pas les frameworks (LangChain, Amazon Bedrock AI Agent etc) le fameux débat sur l'abstraction Beaucoup de patterns implementable avec quelques lignes sans frameworks Plusieurs blocks de complexité croissante Augmented LLM (RAG, memory etc): Anthropic dit que les LLMs savent coordonner cela via MCP apr exemple Second: workflow prompt chaining : avec des gates et appelle les LLMs savent coordonner successivement ; favorise la precision vs la latence vu que les taches sont décomposées en plusieurs calls LLMs Workflow routing: classifie une entree et choisie la route a meilleure: separation de responsabilité Workflow : parallelisation: LLM travaillent en paralllele sur une tache et un aggregateur fait la synthèse. Paralleisaiton avec saucissonage de la tache ou voter sur le meilleur réponse Workflow : orchestrator workers: quand les taches ne sont pas bounded ou connues (genre le nombre de fichiers de code à changer) - les sous taches ne sont pas prédéfinies Workflow: evaluator optimizer: nun LLM propose une réponse, un LLM l'évalue et demande une meilleure réponse au besoin Agents: commande ou interaction avec l;humain puis autonome meme si il peut revenir demander des precisions à l'humain. Agents sont souvent des LLM utilisât des outil pour modifier l'environnement et réagir a feedback en boucle Ideal pour les problèmes ouverts et ou le nombre d'étapes n'est pas connu Recommende d'y aller avec une complexité progressive L'IA c'est pas donné https://techcrunch.com/2025/01/05/openai-is-losing-money-on-its-pricey-chatgpt-pro-plan-ceo-sam-altman-says/ OpenAI annonce que même avec des licenses à 200$/mois ils ne couvrent pas leurs couts associés… A quand l'explosion de la bulle IA ? Outillage Ghostty, un nouveau terminal pour Linux et macOS : https://ghostty.org/ Initié par Mitchell Hashimoto (hashicorp) Ghostty est un émulateur de terminal natif pour macOS et Linux. Il est écrit en Swift et utilise AppKit et SwiftUI sur macOS, et en Zig et utilise l'API GTK4 C sur Linux. Il utilise des composants d'interface utilisateur native et des raccourcis clavier et souris standard. Il prend en charge Quick Look, Force Touch et d'autres fonctionnalités spécifiques à macOS. Ghostty essaie de fournir un ensemble riche de fonctionnalités utiles pour un usage quotidien. Comment Pinterest utilise Honeycomb pour améliorer sa CI https://medium.com/pinterest-engineering/how-pinterest-leverages-honeycomb-to-enhance-ci-observability-and-improve-ci-build-stability–15eede563d75 Pinterest utilise Honeycomb pour améliorer l'observabilité de l'intégration continue (CI). Honeycomb permet à Pinterest de visualiser les métriques de build, d'analyser les tendances et de prendre des décisions basées sur les données. Honeycomb aide également Pinterest à identifier les causes potentielles des échecs de build et à rationaliser les tâches d'astreinte. Honeycomb peut également être utilisé pour suivre les métriques de build locales iOS aux côtés des détails de la machine, ce qui aide Pinterest à prioriser les mises à niveau des ordinateurs portables pour les développeurs. Méthodologies Suite à notre épisode sur les différents types de documentation, cet article parle des bonnes pratiques à suivre pour les tutoriels https://refactoringenglish.com/chapters/rules-for-software-tutorials/ Écrivez des tutoriels pour les débutants, en évitant le jargon et la terminologie complexe. Promettez un résultat clair dans le titre et expliquez l'objectif dans l'introduction. Montrez le résultat final tôt pour réduire les ambiguïtés. Rendez les extraits de code copiables et collables, en évitant les invites de shell et les commandes interactives. Utilisez les versions longues des indicateurs de ligne de commande pour plus de clarté. Séparez les valeurs définies par l'utilisateur de la logique réutilisable à l'aide de variables d'environnement ou de constantes nommées. Épargnez au lecteur les tâches inutiles en utilisant des scripts. Laissez les ordinateurs évaluer la logique conditionnelle, pas le lecteur. Maintenez le code en état de fonctionnement tout au long du tutoriel. Enseignez une chose par tutoriel et minimisez les dépendances. Les Wide events, un “nouveau” concept en observabilité https://jeremymorrell.dev/blog/a-practitioners-guide-to-wide-events/ un autre article https://isburmistrov.substack.com/p/all-you-need-is-wide-events-not-metrics L'idée est de logger des evenements (genre JSON log) avec le plus d'infos possible de la machine, la ram, la versiond e l'appli, l'utilisateur, le numero de build qui a produit l'appli, la derniere PR etc etc ca permet de filtrer et grouper by et de voir des correlations visuelles tres rapidement et de zoomer tiens les ventes baisses de 20% tiens en fait ca vient de l'appli andriod tiens aps correle a la version de l'appli mais la version de l'os si! le deuxieme article est facile a lire le premier est un guide d'usage exhaustif du concept Entre argumenter et se donner 5 minutes https://signalvnoise.com/posts/3124-give-it-five-minutes on veut souvent argumenter aka poser des questions en ayant déjà la reponse en soi emotionnellement mais ca amene beaucoup de verbiage donner 5 minutes à l'idée le temps d'y penser avant d'argumenter Loi, société et organisation Des juges fédéraux arrêtent le principe de la neutralité du net https://www.lemonde.fr/pixels/article/2025/01/03/les-etats-unis-reviennent-en-arriere-sur-le-principe-de-la-neutralite-du-net_6479575_4408996.html?lmd_medium=al&lmd_campaign=envoye-par-appli&lmd_creation=ios&lmd_source=default la neutralité du net c'est l'interdiction de traiter un paquet différemment en fonction de son émetteur Par exemple un paquet Netflix qui serait ralenti vs un paquet Amazon Donald trump est contre cette neutralité. À voir les impacts concrets dans un marché moins régulé. Rubrique débutant Un petit article sur les float vs les double en Java https://www.baeldung.com/java-float-vs-double 4 vs 8 bytes precision max de 7 vs 15 echele 10^38 vs 10^308 (ordre de grandeur) perf a peu pret similaire sauf peut etre pour des modeles d'IA qui vont privilegier une taille plus petite parfois attention overflow et les accumulation d'erreurs d'approximation BigDecimal Conférences La liste des conférences provenant de Developers Conferences Agenda/List par Aurélie Vache et contributeurs : 20 janvier 2025 : Elastic{ON} - Paris (France) 22–25 janvier 2025 : SnowCamp 2025 - Grenoble (France) 24–25 janvier 2025 : Agile Games Île-de-France 2025 - Paris (France) 6–7 février 2025 : Touraine Tech - Tours (France) 21 février 2025 : LyonJS 100 - Lyon (France) 28 février 2025 : Paris TS La Conf - Paris (France) 6 mars 2025 : DevCon #24 : 100% IA - Paris (France) 13 mars 2025 : Oracle CloudWorld Tour Paris - Paris (France) 14 mars 2025 : Rust In Paris 2025 - Paris (France) 19–21 mars 2025 : React Paris - Paris (France) 20 mars 2025 : PGDay Paris - Paris (France) 20–21 mars 2025 : Agile Niort - Niort (France) 25 mars 2025 : ParisTestConf - Paris (France) 26–29 mars 2025 : JChateau Unconference 2025 - Cour-Cheverny (France) 27–28 mars 2025 : SymfonyLive Paris 2025 - Paris (France) 28 mars 2025 : DataDays - Lille (France) 28–29 mars 2025 : Agile Games France 2025 - Lille (France) 3 avril 2025 : DotJS - Paris (France) 3 avril 2025 : SoCraTes Rennes 2025 - Rennes (France) 4 avril 2025 : Flutter Connection 2025 - Paris (France) 10–11 avril 2025 : Android Makers - Montrouge (France) 10–12 avril 2025 : Devoxx Greece - Athens (Greece) 16–18 avril 2025 : Devoxx France - Paris (France) 23–25 avril 2025 : MODERN ENDPOINT MANAGEMENT EMEA SUMMIT 2025 - Paris (France) 24 avril 2025 : IA Data Day 2025 - Strasbourg (France) 29–30 avril 2025 : MixIT - Lyon (France) 7–9 mai 2025 : Devoxx UK - London (UK) 15 mai 2025 : Cloud Toulouse - Toulouse (France) 16 mai 2025 : AFUP Day 2025 Lille - Lille (France) 16 mai 2025 : AFUP Day 2025 Lyon - Lyon (France) 16 mai 2025 : AFUP Day 2025 Poitiers - Poitiers (France) 24 mai 2025 : Polycloud - Montpellier (France) 5–6 juin 2025 : AlpesCraft - Grenoble (France) 5–6 juin 2025 : Devquest 2025 - Niort (France) 11–13 juin 2025 : Devoxx Poland - Krakow (Poland) 12–13 juin 2025 : Agile Tour Toulouse - Toulouse (France) 12–13 juin 2025 : DevLille - Lille (France) 17 juin 2025 : Mobilis In Mobile - Nantes (France) 24 juin 2025 : WAX 2025 - Aix-en-Provence (France) 25–27 juin 2025 : BreizhCamp 2025 - Rennes (France) 26–27 juin 2025 : Sunny Tech - Montpellier (France) 1–4 juillet 2025 : Open edX Conference - 2025 - Palaiseau (France) 7–9 juillet 2025 : Riviera DEV 2025 - Sophia Antipolis (France) 18–19 septembre 2025 : API Platform Conference - Lille (France) & Online 2–3 octobre 2025 : Volcamp - Clermont-Ferrand (France) 6–10 octobre 2025 : Devoxx Belgium - Antwerp (Belgium) 9–10 octobre 2025 : Forum PHP 2025 - Marne-la-Vallée (France) 16–17 octobre 2025 : DevFest Nantes - Nantes (France) 4–7 novembre 2025 : NewCrafts 2025 - Paris (France) 6 novembre 2025 : dotAI 2025 - Paris (France) 7 novembre 2025 : BDX I/O - Bordeaux (France) 12–14 novembre 2025 : Devoxx Morocco - Marrakech (Morocco) 23–25 avril 2026 : Devoxx Greece - Athens (Greece) 17 juin 2026 : Devoxx Poland - Krakow (Poland) Nous contacter Pour réagir à cet épisode, venez discuter sur le groupe Google https://groups.google.com/group/lescastcodeurs Contactez-nous via X/twitter https://twitter.com/lescastcodeurs ou Bluesky https://bsky.app/profile/lescastcodeurs.com Faire un crowdcast ou une crowdquestion Soutenez Les Cast Codeurs sur Patreon https://www.patreon.com/LesCastCodeurs Tous les épisodes et toutes les infos sur https://lescastcodeurs.com/

Go To Market Grit
#225 CEO Lattice, Sarah Franklin: Trailblazer

Go To Market Grit

Play Episode Listen Later Jan 13, 2025 72:01


Guest: Sarah Franklin, CEO of LatticeAs the CEO of a growing company, Lattice's Sarah Franklin has learned that one of her most important contributions is taking a leap of faith. “You have to have the courage to be the first one to do it,” she says,” and to show that it can be done, and to pave the way so that then your team feels trust.”Sarah cautions, though, that sometimes courage is deciding to stop and go a different direction. As agentic AI becomes more common, the people building companies like Lattice should look to the “cautionary tales” of how social media and mobile phones have changed society, she says.“We can have the courage to say, what are the outcomes that we want to prevent? Or what are the outcomes that we want to make sure happen? This all takes, courage, because it's all unknown.”Chapters:(01:14) - Schooling in Mexico (04:09) - Raising brave children (10:28) - Sarah's upbringing (13:29) - The pursuit of money (16:23) - Measuring success (19:28) - Learnings, not regrets (22:55) - Make an impact (26:44) - Pitching Trailhead (32:56) - Elevating a B2B company (35:27) - How to colonize Mars (38:39) - Marketing, the Salesforce way (44:21) - Dolphining and truth-tellers (50:56) - Renewed purpose (56:30) - The challenges of being CEO (01:00:18) - Pave the way (01:03:25) - “Humanizing AI” (01:06:57) - Handling controversy (01:11:04) - Who Lattice is hiring and what “grit” means to Sarah Mentioned in this episode: FaceTime, Salesforce, Marc Benioff, Mahatma Gandhi, Instagram, the Fortune 500, Java, Jerry Maguire, National Parks, Nike, Michael Jordan, Apple and “Think Different,” Sara Varni, Scott Holden, Andy Kofoid, Databricks, Datadog, Behind the Cloud, Oracle, Microsoft, Elon Musk, Amazon AWS, George Hu, Mike Rosenbaum, Cheryl Feldman, Zac Otero, Guidewire Software, AI agents, Indiana Jones and the Last Crusade, and LinkedIn.Links:Connect with SarahLinkedInConnect with JoubinTwitterLinkedInEmail: grit@kleinerperkins.com Learn more about Kleiner PerkinsThis episode was edited by Eric Johnson from LightningPod.fm

AWS for Software Companies Podcast
Ep072: From Alerts to Action - How Datadog Manages Security Incidents with AI

AWS for Software Companies Podcast

Play Episode Listen Later Dec 30, 2024 23:44


Dr. Yanbing Li, Chief Product Officer at Datadog, outlines how the company has integrated AI and automation into its incident response framework, helping customers manage both traditional security challenges and emerging AI-specific risks.Topics Include:Introduced talk about incident response and CISO liabilityDatadog founded 14 years ago for cloud-based developmentPlatform unifies observability and security for cloud applicationsCurrent environment has too many fragmented security productsSEC requires material incident reporting within four daysDatadog's incident response automates Slack room creationResponse team includes Legal, Security, Engineering, and ProductSystem tracks non-material incidents to identify concerning patternsReal-time telemetry data drives incident management automationOn-call capabilities manage escalation workflowsDatadog uses own products internally for incident responseCompany focuses on reducing time to incident detectionAI brings new risks: hallucination, data leaks, design exploitationBits.ai launched as LLM-based incident management co-pilotTool synthesizes events and generates incident summariesBits.ai suggests code remediation and creates synthetic testsSecurity built into AI products from initial designPrompt injection prevented through structured validation approachSensitive data anonymized before LLM processingEngineering and security teams collaborate closely on AILLM observability becoming critical for production deploymentsCustomers need monitoring for hallucinations and token usageDatadog extends infrastructure monitoring into security naturallyCompany maintains strong partnership with AWSQ&A covered Bits.ai proactive capabilities and enterprise differentiationParticipants:Yanbing Li – Chief Produce Officer - DatadogSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon/isv/

Screaming in the Cloud
Replay - Hacking AWS in Good Faith with Nick Frichette

Screaming in the Cloud

Play Episode Listen Later Dec 26, 2024 32:32


On this Screaming in the Cloud Replay, we're taking you back to our chat with Nick Frichette. He's the maintainer of hackingthe.cloud, and holds security and solutions architect AWS certifications, and in his spare time, he conducts vulnerability research at Hacking the Cloud. Join Corey and Nick as they talk about the various kinds of cloud security researchers and touch upon offensive security, why Nick decided to create Hacking the Cloud, how AWS lets security researchers conduct penetration testing in good faith, some of the more interesting AWS exploits Nick has discovered, how it's fun to play keep-away with incident response, why you need to get legal approval before conducting penetration testing, and more.Show Highlights(0:00) Intro(0:42) The Duckbill Group sponsor read(1:15) What is a Cloud Security Researcher?(3:49) Nick's work with Hacking the Cloud(5:24) Building relationships with cloud providers(7:34) Nick's security findings through cloud logs(13:05) How Nick finds security flaws(15:31) Reporting vulnerabilities to AWS and “bug bounty” programs(18:41) The Duckbill Group sponsor read(19:24) How to report vulnerabilities ethically(21:52) Good disclosure programs vs. bad ones(28:23) What's next for Nick(31:27) Where you can find more from NickAbout Nick FrichetteNick Frichette is a Staff Security Researcher at Datadog, specializing in offensive security within AWS environments. His focus is on discovering new attack vectors targeting AWS services, environments, and applications. From his research, Nick develops detection methods and preventive measures to secure these systems. Nick's work often leads to the discovery of vulnerabilities within AWS itself, and he collaborates closely with Amazon to ensure they are remediated.Nick has also presented his research at major industry conferences, including Black Hat USA, DEF CON, fwd:cloudsec, and others.LinksHacking the Cloud: https://hackingthe.cloud/Determine the account ID that owned an S3 bucket vulnerability: https://hackingthe.cloud/aws/enumeration/account_id_from_s3_bucket/Twitter: https://twitter.com/frichette_nPersonal website:https://frichetten.comOriginal Episodehttps://www.lastweekinaws.com/podcast/screaming-in-the-cloud/hacking-aws-in-good-faith-with-nick-frichette/SponsorThe Duckbill Group: duckbillgroup.com

GOTO - Today, Tomorrow and the Future
Observability 2.0: Transforming Logging & Metrics • Charity Majors & James Lewis

GOTO - Today, Tomorrow and the Future

Play Episode Listen Later Dec 20, 2024 30:35 Transcription Available


This interview was recorded for GOTO Unscripted.https://gotopia.techRead the full transcription of this interview here:https://gotopia.tech/articles/336Charity Majors - CTO at honeycomb.io & Co- Author of "Observability Engineering"James Lewis - Software Architect & Director at ThoughtworksRESOURCESCharityhttps://twitter.com/mipsytipsyhttps://github.com/charityhttps://linkedin.com/in/charity-majorshttps://charity.wtfJameshttps://twitter.com/boicyhttps://linkedin.com/in/james-lewis-microserviceshttps://github.com/boicyhttps://www.bovon.orgDESCRIPTIONWhat's next in the observability space? Join Charity Majors and James Lewis as they discuss canonical logs, Observability 2.0 and how to simplify complexity in software engineering!RECOMMENDED BOOKSCharity Majors, Liz Fong-Jones & George Miranda • Observability Engineering • https://amzn.to/38scbmaCharity Majors & Laine Campbell • Database Reliability Engineering • https://amzn.to/3ujybdSKelly Shortridge & Aaron Rinehart • Security Chaos Engineering • https://www.verica.io/sce-bookNora Jones & Casey Rosenthal • Chaos Engineering • https://amzn.to/3hUmuAHRuss Miles • Learning Chaos Engineering • https://amzn.to/3hCiUe8Nicole Forsgren, Jez Humble & Gene Kim • Accelerate • https://amzn.to/442Rep0BlueskyTwitterInstagramLinkedInFacebookCHANNEL MEMBERSHIP BONUSJoin this channel to get early access to videos & other perks:https://www.youtube.com/channel/UCs_tLP3AiwYKwdUHpltJPuA/joinLooking for a unique learning experience?Attend the next GOTO conference near you! Get your ticket: gotopia.techSUBSCRIBE TO OUR YOUTUBE CHANNEL - new videos posted daily!

Google Cloud Cast
Saiba o que o Google Cloud Marketplace pode fazer pelo seu negócio

Google Cloud Cast

Play Episode Listen Later Dec 5, 2024 36:37


Hoje, as empresas compram, testam, validam, implementam e administram soluções de tecnologia de uma forma bem diferente do que era feito no passado. Não só os negócios levaram suas infraestruturas para a nuvem, mas a nuvem também chegou a todos os pontos de contato do mercado, inclusive na criação de aplicações, no uso de softwares e no próprio processo de compra dessas soluções. Assim, nasceram também os marketplaces que vieram como uma plataforma para unir serviços e produtos tecnológicos em um mesmo lugar, além de serem uma resposta a essa mudança sobre como empresas hoje adquirem suas soluções. Mas como é que funciona isso? Daniel Leite, Especialista de Vendas do Google Cloud, e Marcelo Gomes, Especialista de Vendas para IA do Google Cloud, recebem Renata Faria, Head de Parceiros ISVs na América Latina, e Everton Ribeiro, Gerente de Parcerias da Datadog, para falar sobre Google Cloud Marketplace. Gostou do episódio? Tem alguma dúvida ou sugestão? Fale com a gente pelo e-mail googlecloudcast@google.com.

The Twenty Minute VC: Venture Capital | Startup Funding | The Pitch
20Sales: What I Learned Scaling Datadog from $60M to $1BN in ARR | How to do Outbound in 2024 | Why Discounting is Dangerous and Contract Sizes are Misleading with Dan Fougere

The Twenty Minute VC: Venture Capital | Startup Funding | The Pitch

Play Episode Listen Later Nov 22, 2024 70:05


Dan Fougere is one of the most successful sales leaders of the last decade. Most recently, Dan was Chief Revenue Officer for Datadog, growing revenues from $60 million to $1BN ARR. Before Datadog, Dan was Head of Global Sales at Medallia where he created the Mediallia sales playbook. In addition, Dan is also a minority owner of the New York Yankees.  In Today's Episode with Dan Fougere:  1. Lessons Scaling Sales to $1BN in ARR at Datadog: What did Datadog not do that Dan wishes they had of done? What did they not do that Dan wishes they had done? What does Dan know about scaling sales to $1BN in ARR that he wishes he had known at the beginning? What stage of the scaling process was hardest? Why? 2. How to Hire the Best Sales Team: What are the top signals of the best sales candidates? How does Dan structure the interview process for new candidates? How does Dan use tasks and take-home assignments to test candidates? What does Dan think of hiring panels? What are the biggest hiring mistakes Dan has made? What did he learn? 3. Discounting, Logos and Deal Reviews: Is discounting always wrong? How should sales leaders use it? How important is the quality of logo in the early days vs revenue in the door? What is the right way to structure deal reviews? What makes good vs great? Is outbound dead in 2024? Advice to founders on outbound?  

AWS Morning Brief
The Return of Old AWS

AWS Morning Brief

Play Episode Listen Later Nov 18, 2024 4:53


AWS Morning Brief for the week of November 18, with Corey Quinn. Links:Buy a shirt benefiting 826 National!Amazon DataZone updates pricing and removes the user-level subscription feeAmazon DynamoDB reduces prices for on-demand throughput and global tablesAmazon DynamoDB introduces warm throughput for tables and indexesAmazon EBS now supports detailed performance statistics on EBS volume healthAmazon Q Developer plugins for Datadog and Wiz now generally availableAmazon S3 now supports up to 1 million buckets per AWS accountAWS Backup now supports copying Amazon S3 backups across Regions and accounts in opt-in RegionsAWS CloudTrail Lake announces enhanced event filteringHow and why you should move to Cost and Usage Report (CUR) 2.0?AWS BuilderCards second edition at re:Invent 2024Accelerate your third-party Amazon EKS add-on onboarding using ConformitronPython 3.13 runtime now available in AWS LambdaDeploy the Cost Optimizer for Amazon WorkSpaces in a highly-regulated environment.Introducing the Live Event Framework: Live Streaming with Ad Insertion on AWSIntroducing kro: Kube Resource OrchestratorAWS Snow device updates

Ultimate Guide to Partnering™
243 – Customer-Centric Strategies in Google Cloud Marketplace: Insight from Datadog

Ultimate Guide to Partnering™

Play Episode Listen Later Nov 12, 2024 7:57


I am excited to bring you insights from my conversation with Jarrod Buckley, VP of Channels and Alliances at Datadog, live from Google Cloud's Marketplace Exchange! Datadog, a leader in cloud application monitoring and security, is transforming how customers manage their cloud environments through a single, unified platform. In this episode, Jarrod shares how Datadog leverages Google Cloud Marketplace to simplify customer procurement and drive larger, faster deals. He highlights Datadog's recent innovations, including empowering channel partners to sell Datadog through Google Cloud Marketplace and the ecosystem approach that places customer needs at the center of co-selling strategies. As marketplace adoption grows, Datadog is scaling across geographies, with systems and human teams aligning to ensure seamless co-selling at scale. For anyone looking to optimize their marketplace strategy, Jarrod offers valuable advice for planning and aligning incentives as we head into 2025. Catch this episode of The Ultimate Guide to Partnering for insights into how Datadog leads the charge in marketplace-driven cloud success!

#plugintodevin - Your Mark on the World with Devin Thorpe
Tackling the Food Waste Crisis with Robotics: How ZeroDay's Tech Revolutionizes Urban Sustainability

#plugintodevin - Your Mark on the World with Devin Thorpe

Play Episode Listen Later Nov 12, 2024 24:24


I'm not a financial advisor; Superpowers for Good should not be considered investment advice. Seek counsel before making investment decisions.Watch the show on television by downloading the e360tv channel app to your Roku, AppleTV or AmazonFireTV. You can also see it on YouTube.When you purchase an item, launch a campaign or create an investment account after clicking a link here, we may earn a fee. Engage to support our work.Devin: What is your superpower?Maneesa: I would say my superpower is an inability to give up.Food waste is the largest contributor to the municipal waste stream, yet it's managed with century-old methods like trash chutes and compactors. This is where ZeroDay, under the leadership of CEO and co-founder Maneesa Wijesinghe-Nelken, steps in. ZeroDay's solution: robotics and automation designed to handle food waste at the source, within urban buildings like hotels, offices, and restaurants. This innovative approach offers both a reduction in labor and a decrease in the operational expenses associated with traditional waste management.ZeroDay's technology compresses food waste into manageable blocks wrapped in a special bio-wrapper. “Our machines achieve significant volume reduction by up to 75 percent,” Maneesa explains, “which cuts down on the truck trips needed to collect and transport waste, ultimately lowering costs for composting or biogas production.”The potential impact of this innovation is immense. Food waste in landfills emits methane—a greenhouse gas approximately 80 times more potent than CO₂ over a 20-year period. By diverting food waste from landfills and into sustainable processes like composting and biogas production, ZeroDay contributes to reducing urban carbon footprints.Based in New York City, where waste management challenges are pervasive, ZeroDay is piloting its technology in an environment ripe for change. Maneesa envisions scaling this model to other major cities, where food waste management remains a pressing issue.tl;dr:* Today's episode highlights ZeroDay's robotic solution that compresses food waste into sustainable, compact blocks.* Maneesa explains how ZeroDay's technology reduces landfill waste and methane emissions significantly.* ZeroDay's automation helps urban buildings manage food waste more efficiently, saving labor and operational costs.* Maneesa's unshakeable optimism drives her to overcome challenges and create impactful technology solutions.* ZeroDay is currently piloting its waste management technology with business customers across New York City.How to Develop Relentless Determination As a SuperpowerManeesa's superpower is a relentless determination to succeed. She describes herself as “a blind optimist” who refuses to give up, driven by an unwavering belief that she'll overcome any obstacles. This internal drive allows her to push through challenges and keep moving forward, even when the path isn't clear.An example of Maneesa's relentless determination is her experience building ZeroDay's first working prototype. Despite limited experience in hard tech, she and her co-founder faced the intense challenge of building a robotic food-waste processing machine in a month for a demonstration. Maneesa's commitment was so deep that, even after injuring herself on a circular saw, she persisted and completed the prototype. Their efforts led to a successful demo that generated significant interest from businesses and validated their concept despite the grueling process.Tips for Developing This Superpower:* Cultivate blind optimism by finding an inner voice that reminds you success is possible.* Embrace spirituality or meditation practices to build resilience and maintain focus.* Approach challenges with an open mind, seeing each one as an opportunity to learn.* Believe in the purpose behind your work to fuel your drive, even in tough times.Closing Paragraph:By following Maneesa's example and advice, you can make relentless determination a skill. With practice and effort, you could make it a superpower that enables you to do more good in the world.Remember, however, that research into success suggests that building on your own superpowers is more important than creating new ones or overcoming weaknesses. You do you!Guest ProfileManeesa Wijesinghe-Nelken (she/her):CEO & Co-founder, ZeroDayAbout ZeroDay: ZeroDay automates the archaic food waste management system in businesses, allowing them to save money, comply with regulations, and prevent food waste from entering landfills.Website: www.zeroday.lifeBiographical Information: Maneesa previously served as the Director of the Regenerative and Circular Economy Lab, a think tank at the University of Oxford's Smith School of Enterprise and Environment. In this role, she collaborated with large corporations such as Nike and Unilever and small local organizations to develop and implement circular business solutions on various scales. She is also a fellow of the Ellen MacArthur Foundation and has completed a study to earn the "From Linear to Circular" professional certification. She holds an MBA from the University of Oxford, specializing in circular business models. Before this, she was an early employee at disruptive tech companies such as Datadog and Okta, focusing on technical solutions and enterprise sales. Maneesa's educational background includes a Bachelors in Physics and Electrical Engineering from Northwestern University. As a native of Sri Lanka, she has firsthand experience of the environmental damage caused by the waste crisis, including witnessing the 2017 Meethotamulla garbage dump disaster. Her experiences have fueled her ambition to combat the waste crisis.LinkedIn: linkedin.com/in/maneesawn/Support Our SponsorsOur generous sponsors make our work possible, serving impact investors, social entrepreneurs, community builders and diverse founders. Today's advertisers include FundingHope, Mivium, Abby and The SuperCrowd Mastermind Group. Learn more about advertising with us here.Max-Impact MembersThe following Max-Impact Members provide valuable financial support to keep us operating:Carol Fineagan, Independent Consultant | Lory Moore, Lory Moore Law | Marcia Brinton, High Desert Gear | Paul Lovejoy, Stakeholder Enterprise | Ralf Mandt, Next Pitch | Add Your Name HereUpcoming SuperCrowd Event CalendarIf a location is not noted, the events below are virtual.* SuperCrowd Mastermind Group, twice monthly on the 1st and 3rd Thursdays at noon Eastern. This group is for entrepreneurs and small business owners interested in raising money from the crowd. Attend your first meeting free!* Superpowers for Good Televised Live Pitch, November 13, 9:00 PM Eastern during primetime. At the event, judges will select their pick, and the audience will select the SuperCrowd Award recipient. Put the date on your calendar to watch it live!* Impact Cherub Club Meeting hosted by The Super Crowd, Inc., a public benefit corporation, on November 19, 2024, at 1:00 PM Eastern. Each month, the Club meets to review new offerings for investment consideration and to conduct due diligence on previously screened deals. To join the Impact Cherub Club, become an Impact Member of the SuperCrowd.* SuperCrowdHour, November 20, 2024, at 1:00 PM Eastern. Each month, we host a value-laden webinar for aspiring impact investors or social entrepreneurs. At November's SuperCrowdHour, Devin will explain six common investment types you need to understand before you can invest like a pro. Free to attend.Community Event Calendar* Successful Funding with Karl Dakin, Tuesdays at 10:00 AM ET - Click on Events* Community Revitalization, Thursdays, 10:00 AM Eastern.* Crowdfunding Unlocked, WorldTree, Thursday, November 14 at 3:00 ET/Noon PT. Devin Thorpe, Léa Bouhelier-Gautreau and Cathy Key.* Main Street Skowhegan and NC3 Entrepreneur Finance Workshop Series, September 17 - November 19, 2023.* 2025 Earthshot Prize Application window open through November 15, 2024. Apply today!* Asheville Neighborhood Economics, date TBD following impact of Helene.If you would like to submit an event for us to share with the 8,000+ members of the SuperCrowd, click here.We use AI to help us write compelling recaps of each episode. Get full access to Superpowers for Good at www.superpowers4good.com/subscribe

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

We are recording our next big recap episode and taking questions! Submit questions and messages on Speakpipe here for a chance to appear on the show!Also subscribe to our calendar for our Singapore, NeurIPS, and all upcoming meetups!In our first ever episode with Logan Kilpatrick we called out the two hottest LLM frameworks at the time: LangChain and Dust. We've had Harrison from LangChain on twice (as a guest and as a co-host), and we've now finally come full circle as Stanislas from Dust joined us in the studio.After stints at Oracle and Stripe, Stan had joined OpenAI to work on mathematical reasoning capabilities. He describes his time at OpenAI as "the PhD I always wanted to do" while acknowledging the challenges of research work: "You're digging into a field all day long for weeks and weeks, and you find something, you get super excited for 12 seconds. And at the 13 seconds, you're like, 'oh, yeah, that was obvious.' And you go back to digging." This experience, combined with early access to GPT-4's capabilities, shaped his decision to start Dust: "If we believe in AGI and if we believe the timelines might not be too long, it's actually the last train leaving the station to start a company. After that, it's going to be computers all the way down."The History of DustDust's journey can be broken down into three phases:* Developer Framework (2022): Initially positioned as a competitor to LangChain, Dust started as a developer tooling platform. While both were open source, their approaches differed – LangChain focused on broad community adoption and integration as a pure developer experience, while Dust emphasized UI-driven development and better observability that wasn't just `print` statements.* Browser Extension (Early 2023): The company pivoted to building XP1, a browser extension that could interact with web content. This experiment helped validate user interaction patterns with AI, even while using less capable models than GPT-4.* Enterprise Platform (Current): Today, Dust has evolved into an infrastructure platform for deploying AI agents within companies, with impressive metrics like 88% daily active users in some deployments.The Case for Being HorizontalThe big discussion for early stage companies today is whether or not to be horizontal or vertical. Since models are so good at general tasks, a lot of companies are building vertical products that take care of a workflow end-to-end in order to offer more value and becoming more of “Services as Software”. Dust on the other hand is a platform for the users to build their own experiences, which has had a few advantages:* Maximum Penetration: Dust reports 60-70% weekly active users across entire companies, demonstrating the potential reach of horizontal solutions rather than selling into a single team.* Emergent Use Cases: By allowing non-technical users to create agents, Dust enables use cases to emerge organically from actual business needs rather than prescribed solutions.* Infrastructure Value: The platform approach creates lasting value through maintained integrations and connections, similar to how Stripe's value lies in maintaining payment infrastructure. Rather than relying on third-party integration providers, Dust maintains its own connections to ensure proper handling of different data types and structures.The Vertical ChallengeHowever, this approach comes with trade-offs:* Harder Go-to-Market: As Stan talked about: "We spike at penetration... but it makes our go-to-market much harder. Vertical solutions have a go-to-market that is much easier because they're like, 'oh, I'm going to solve the lawyer stuff.'"* Complex Infrastructure: Building a horizontal platform requires maintaining numerous integrations and handling diverse data types appropriately – from structured Salesforce data to unstructured Notion pages. As you scale integrations, the cost of maintaining them also scales. * Product Surface Complexity: Creating an interface that's both powerful and accessible to non-technical users requires careful design decisions, down to avoiding technical terms like "system prompt" in favor of "instructions." The Future of AI PlatformsStan initially predicted we'd see the first billion-dollar single-person company in 2023 (a prediction later echoed by Sam Altman), but he's now more focused on a different milestone: billion-dollar companies with engineering teams of just 20 people, enabled by AI assistance.This vision aligns with Dust's horizontal platform approach – building the infrastructure that allows small teams to achieve outsized impact through AI augmentation. Rather than replacing entire job functions (the vertical approach), they're betting on augmenting existing workflows across organizations.Full YouTube EpisodeChapters* 00:00:00 Introductions* 00:04:33 Joining OpenAI from Paris* 00:09:54 Research evolution and compute allocation at OpenAI* 00:13:12 Working with Ilya Sutskever and OpenAI's vision* 00:15:51 Leaving OpenAI to start Dust* 00:18:15 Early focus on browser extension and WebGPT-like functionality* 00:20:20 Dust as the infrastructure for agents* 00:24:03 Challenges of building with early AI models* 00:28:17 LLMs and Workflow Automation* 00:35:28 Building dependency graphs of agents* 00:37:34 Simulating API endpoints* 00:40:41 State of AI models* 00:43:19 Running evals* 00:46:36 Challenges in building AI agents infra* 00:49:21 Buy vs. build decisions for infrastructure components* 00:51:02 Future of SaaS and AI's Impact on Software* 00:53:07 The single employee $1B company race* 00:56:32 Horizontal vs. vertical approaches to AI agentsTranscriptAlessio [00:00:00]: Hey everyone, welcome to the Latent Space podcast. This is Alessio, partner and CTO at Decibel Partners, and I'm joined by my co-host Swyx, founder of Smol.ai.Swyx [00:00:11]: Hey, and today we're in a studio with Stanislas, welcome.Stan [00:00:14]: Thank you very much for having me.Swyx [00:00:16]: Visiting from Paris.Stan [00:00:17]: Paris.Swyx [00:00:18]: And you have had a very distinguished career. It's very hard to summarize, but you went to college in both Ecopolytechnique and Stanford, and then you worked in a number of places, Oracle, Totems, Stripe, and then OpenAI pre-ChatGPT. We'll talk, we'll spend a little bit of time about that. About two years ago, you left OpenAI to start Dust. I think you were one of the first OpenAI alum founders.Stan [00:00:40]: Yeah, I think it was about at the same time as the Adept guys, so that first wave.Swyx [00:00:46]: Yeah, and people really loved our David episode. We love a few sort of OpenAI stories, you know, for back in the day, like we're talking about pre-recording. Probably the statute of limitations on some of those stories has expired, so you can talk a little bit more freely without them coming after you. But maybe we'll just talk about, like, what was your journey into AI? You know, you were at Stripe for almost five years, there are a lot of Stripe alums going into OpenAI. I think the Stripe culture has come into OpenAI quite a bit.Stan [00:01:11]: Yeah, so I think the buses of Stripe people really started flowing in, I guess, after ChatGPT. But, yeah, my journey into AI is a... I mean, Greg Brockman. Yeah, yeah. From Greg, of course. And Daniela, actually, back in the days, Daniela Amodei.Swyx [00:01:27]: Yes, she was COO, I mean, she is COO, yeah. She had a pretty high job at OpenAI at the time, yeah, for sure.Stan [00:01:34]: My journey started as anybody else, you're fascinated with computer science and you want to make them think, it's awesome, but it doesn't work. I mean, it was a long time ago, it was like maybe 16, so it was 25 years ago. Then the first big exposure to AI would be at Stanford, and I'm going to, like, disclose a whole lamb, because at the time it was a class taught by Andrew Ng, and there was no deep learning. It was half features for vision and a star algorithm. So it was fun. But it was the early days of deep learning. At the time, I think a few years after, it was the first project at Google. But you know, that cat face or the human face trained from many images. I went to, hesitated doing a PhD, more in systems, eventually decided to go into getting a job. Went at Oracle, started a company, did a gazillion mistakes, got acquired by Stripe, worked with Greg Buckman there. And at the end of Stripe, I started interesting myself in AI again, felt like it was the time, you had the Atari games, you had the self-driving craziness at the time. And I started exploring projects, it felt like the Atari games were incredible, but there were still games. And I was looking into exploring projects that would have an impact on the world. And so I decided to explore three things, self-driving cars, cybersecurity and AI, and math and AI. It's like I sing it by a decreasing order of impact on the world, I guess.Swyx [00:03:01]: Discovering new math would be very foundational.Stan [00:03:03]: It is extremely foundational, but it's not as direct as driving people around.Swyx [00:03:07]: Sorry, you're doing this at Stripe, you're like thinking about your next move.Stan [00:03:09]: No, it was at Stripe, kind of a bit of time where I started exploring. I did a bunch of work with friends on trying to get RC cars to drive autonomously. Almost started a company in France or Europe about self-driving trucks. We decided to not go for it because it was probably very operational. And I think the idea of the company, of the team wasn't there. And also I realized that if I wake up a day and because of a bug I wrote, I killed a family, it would be a bad experience. And so I just decided like, no, that's just too crazy. And then I explored cybersecurity with a friend. We're trying to apply transformers to cut fuzzing. So cut fuzzing, you have kind of an algorithm that goes really fast and tries to mutate the inputs of a library to find bugs. And we tried to apply a transformer to that and do reinforcement learning with the signal of how much you propagate within the binary. Didn't work at all because the transformers are so slow compared to evolutionary algorithms that it kind of didn't work. Then I started interested in math and AI and started working on SAT solving with AI. And at the same time, OpenAI was kind of starting the reasoning team that were tackling that project as well. I was in touch with Greg and eventually got in touch with Ilya and finally found my way to OpenAI. I don't know how much you want to dig into that. The way to find your way to OpenAI when you're in Paris was kind of an interesting adventure as well.Swyx [00:04:33]: Please. And I want to note, this was a two-month journey. You did all this in two months.Stan [00:04:38]: The search.Swyx [00:04:40]: Your search for your next thing, because you left in July 2019 and then you joined OpenAI in September.Stan [00:04:45]: I'm going to be ashamed to say that.Swyx [00:04:47]: You were searching before. I was searching before.Stan [00:04:49]: I mean, it's normal. No, the truth is that I moved back to Paris through Stripe and I just felt the hardship of being remote from your team nine hours away. And so it kind of freed a bit of time for me to start the exploration before. Sorry, Patrick. Sorry, John.Swyx [00:05:05]: Hopefully they're listening. So you joined OpenAI from Paris and from like, obviously you had worked with Greg, but notStan [00:05:13]: anyone else. No. Yeah. So I had worked with Greg, but not Ilya, but I had started chatting with Ilya and Ilya was kind of excited because he knew that I was a good engineer through Greg, I presume, but I was not a trained researcher, didn't do a PhD, never did research. And I started chatting and he was excited all the way to the point where he was like, hey, come pass interviews, it's going to be fun. I think he didn't care where I was, he just wanted to try working together. So I go to SF, go through the interview process, get an offer. And so I get Bob McGrew on the phone for the first time, he's like, hey, Stan, it's awesome. You've got an offer. When are you coming to SF? I'm like, hey, it's awesome. I'm not coming to the SF. I'm based in Paris and we just moved. He was like, hey, it's awesome. Well, you don't have an offer anymore. Oh, my God. No, it wasn't as hard as that. But that's basically the idea. And it took me like maybe a couple more time to keep chatting and they eventually decided to try a contractor set up. And that's how I kind of started working at OpenAI, officially as a contractor, but in practice really felt like being an employee.Swyx [00:06:14]: What did you work on?Stan [00:06:15]: So it was solely focused on math and AI. And in particular in the application, so the study of the larger grid models, mathematical reasoning capabilities, and in particular in the context of formal mathematics. The motivation was simple, transformers are very creative, but yet they do mistakes. Formal math systems are of the ability to verify a proof and the tactics they can use to solve problems are very mechanical, so you miss the creativity. And so the idea was to try to explore both together. You would get the creativity of the LLMs and the kind of verification capabilities of the formal system. A formal system, just to give a little bit of context, is a system in which a proof is a program and the formal system is a type system, a type system that is so evolved that you can verify the program. If the type checks, it means that the program is correct.Swyx [00:07:06]: Is the verification much faster than actually executing the program?Stan [00:07:12]: Verification is instantaneous, basically. So the truth is that what you code in involves tactics that may involve computation to search for solutions. So it's not instantaneous. You do have to do the computation to expand the tactics into the actual proof. The verification of the proof at the very low level is instantaneous.Swyx [00:07:32]: How quickly do you run into like, you know, halting problem PNP type things, like impossibilities where you're just like that?Stan [00:07:39]: I mean, you don't run into it at the time. It was really trying to solve very easy problems. So I think the... Can you give an example of easy? Yeah, so that's the mass benchmark that everybody knows today. The Dan Hendricks one. The Dan Hendricks one, yeah. And I think it was the low end part of the mass benchmark at the time, because that mass benchmark includes AMC problems, AMC 8, AMC 10, 12. So these are the easy ones. Then AIME problems, somewhat harder, and some IMO problems, like Crazy Arm.Swyx [00:08:07]: For our listeners, we covered this in our Benchmarks 101 episode. AMC is literally the grade of like high school, grade 8, grade 10, grade 12. So you can solve this. Just briefly to mention this, because I don't think we'll touch on this again. There's a bit of work with like Lean, and then with, you know, more recently with DeepMind doing like scoring like silver on the IMO. Any commentary on like how math has evolved from your early work to today?Stan [00:08:34]: I mean, that result is mind blowing. I mean, from my perspective, spent three years on that. At the same time, Guillaume Lampe in Paris, we were both in Paris, actually. He was at FAIR, was working on some problems. We were pushing the boundaries, and the goal was the IMO. And we cracked a few problems here and there. But the idea of getting a medal at an IMO was like just remote. So this is an impressive result. And we can, I think the DeepMind team just did a good job of scaling. I think there's nothing too magical in their approach, even if it hasn't been published. There's a Dan Silver talk from seven days ago where it goes a little bit into more details. It feels like there's nothing magical there. It's really applying reinforcement learning and scaling up the amount of data that can generate through autoformalization. So we can dig into what autoformalization means if you want.Alessio [00:09:26]: Let's talk about the tail end, maybe, of the OpenAI. So you joined, and you're like, I'm going to work on math and do all of these things. I saw on one of your blog posts, you mentioned you fine-tuned over 10,000 models at OpenAI using 10 million A100 hours. How did the research evolve from the GPD 2, and then getting closer to DaVinci 003? And then you left just before ChatGPD was released, but tell people a bit more about the research path that took you there.Stan [00:09:54]: I can give you my perspective of it. I think at OpenAI, there's always been a large chunk of the compute that was reserved to train the GPTs, which makes sense. So it was pre-entropic splits. Most of the compute was going to a product called Nest, which was basically GPT-3. And then you had a bunch of, let's say, remote, not core research teams that were trying to explore maybe more specific problems or maybe the algorithm part of it. The interesting part, I don't know if it was where your question was going, is that in those labs, you're managing researchers. So by definition, you shouldn't be managing them. But in that space, there's a managing tool that is great, which is compute allocation. Basically by managing the compute allocation, you can message the team of where you think the priority should go. And so it was really a question of, you were free as a researcher to work on whatever you wanted. But if it was not aligned with OpenAI mission, and that's fair, you wouldn't get the compute allocation. As it happens, solving math was very much aligned with the direction of OpenAI. And so I was lucky to generally get the compute I needed to make good progress.Swyx [00:11:06]: What do you need to show as incremental results to get funded for further results?Stan [00:11:12]: It's an imperfect process because there's a bit of a... If you're working on math and AI, obviously there's kind of a prior that it's going to be aligned with the company. So it's much easier than to go into something much more risky, much riskier, I guess. You have to show incremental progress, I guess. It's like you ask for a certain amount of compute and you deliver a few weeks after and you demonstrate that you have a progress. Progress might be a positive result. Progress might be a strong negative result. And a strong negative result is actually often much harder to get or much more interesting than a positive result. And then it generally goes into, as any organization, you would have people finding your project or any other project cool and fancy. And so you would have that kind of phase of growing up compute allocation for it all the way to a point. And then maybe you reach an apex and then maybe you go back mostly to zero and restart the process because you're going in a different direction or something else. That's how I felt. Explore, exploit. Yeah, exactly. Exactly. Exactly. It's a reinforcement learning approach.Swyx [00:12:14]: Classic PhD student search process.Alessio [00:12:17]: And you were reporting to Ilya, like the results you were kind of bringing back to him or like what's the structure? It's almost like when you're doing such cutting edge research, you need to report to somebody who is actually really smart to understand that the direction is right.Stan [00:12:29]: So we had a reasoning team, which was working on reasoning, obviously, and so math in general. And that team had a manager, but Ilya was extremely involved in the team as an advisor, I guess. Since he brought me in OpenAI, I was lucky to mostly during the first years to have kind of a direct access to him. He would really coach me as a trainee researcher, I guess, with good engineering skills. And Ilya, I think at OpenAI, he was the one showing the North Star, right? He was his job and I think he really enjoyed it and he did it super well, was going through the teams and saying, this is where we should be going and trying to, you know, flock the different teams together towards an objective.Swyx [00:13:12]: I would say like the public perception of him is that he was the strongest believer in scaling. Oh, yeah. Obviously, he has always pursued the compression thesis. You have worked with him personally, what does the public not know about how he works?Stan [00:13:26]: I think he's really focused on building the vision and communicating the vision within the company, which was extremely useful. I was personally surprised that he spent so much time, you know, working on communicating that vision and getting the teams to work together versus...Swyx [00:13:40]: To be specific, vision is AGI? Oh, yeah.Stan [00:13:42]: Vision is like, yeah, it's the belief in compression and scanning computes. I remember when I started working on the Reasoning team, the excitement was really about scaling the compute around Reasoning and that was really the belief we wanted to ingrain in the team. And that's what has been useful to the team and with the DeepMind results shows that it was the right approach with the success of GPT-4 and stuff shows that it was the right approach.Swyx [00:14:06]: Was it according to the neural scaling laws, the Kaplan paper that was published?Stan [00:14:12]: I think it was before that, because those ones came with GPT-3, basically at the time of GPT-3 being released or being ready internally. But before that, there really was a strong belief in scale. I think it was just the belief that the transformer was a generic enough architecture that you could learn anything. And that was just a question of scaling.Alessio [00:14:33]: Any other fun stories you want to tell? Sam Altman, Greg, you know, anything.Stan [00:14:37]: Weirdly, I didn't work that much with Greg when I was at OpenAI. He had always been mostly focused on training the GPTs and rightfully so. One thing about Sam Altman, he really impressed me because when I joined, he had joined not that long ago and it felt like he was kind of a very high level CEO. And I was mind blown by how deep he was able to go into the subjects within a year or something, all the way to a situation where when I was having lunch by year two, I was at OpenAI with him. He would just quite know deeply what I was doing. With no ML background. Yeah, with no ML background, but I didn't have any either, so I guess that explains why. But I think it's a question about, you don't necessarily need to understand the very technicalities of how things are done, but you need to understand what's the goal and what's being done and what are the recent results and all of that in you. And we could have kind of a very productive discussion. And that really impressed me, given the size at the time of OpenAI, which was not negligible.Swyx [00:15:44]: Yeah. I mean, you've been a, you were a founder before, you're a founder now, and you've seen Sam as a founder. How has he affected you as a founder?Stan [00:15:51]: I think having that capability of changing the scale of your attention in the company, because most of the time you operate at a very high level, but being able to go deep down and being in the known of what's happening on the ground is something that I feel is really enlightening. That's not a place in which I ever was as a founder, because first company, we went all the way to 10 people. Current company, there's 25 of us. So the high level, the sky and the ground are pretty much at the same place. No, you're being too humble.Swyx [00:16:21]: I mean, Stripe was also like a huge rocket ship.Stan [00:16:23]: Stripe, I was a founder. So I was, like at OpenAI, I was really happy being on the ground, pushing the machine, making it work. Yeah.Swyx [00:16:31]: Last OpenAI question. The Anthropic split you mentioned, you were around for that. Very dramatic. David also left around that time, you left. This year, we've also had a similar management shakeup, let's just call it. Can you compare what it was like going through that split during that time? And then like, does that have any similarities now? Like, are we going to see a new Anthropic emerge from these folks that just left?Stan [00:16:54]: That I really, really don't know. At the time, the split was pretty surprising because they had been trying GPT-3, it was a success. And to be completely transparent, I wasn't in the weeds of the splits. What I understood of it is that there was a disagreement of the commercialization of that technology. I think the focal point of that disagreement was the fact that we started working on the API and wanted to make those models available through an API. Is that really the core disagreement? I don't know.Swyx [00:17:25]: Was it safety?Stan [00:17:26]: Was it commercialization?Swyx [00:17:27]: Or did they just want to start a company?Stan [00:17:28]: Exactly. Exactly. That I don't know. But I think what I was surprised of is how quickly OpenAI recovered at the time. And I think it's just because we were mostly a research org and the mission was so clear that some divergence in some teams, some people leave, the mission is still there. We have the compute. We have a site. So it just keeps going.Swyx [00:17:50]: Very deep bench. Like just a lot of talent. Yeah.Alessio [00:17:53]: So that was the OpenAI part of the history. Exactly. So then you leave OpenAI in September 2022. And I would say in Silicon Valley, the two hottest companies at the time were you and Lanktrain. What was that start like and why did you decide to start with a more developer focused kind of like an AI engineer tool rather than going back into some more research and something else?Stan [00:18:15]: Yeah. First, I'm not a trained researcher. So going through OpenAI was really kind of the PhD I always wanted to do. But research is hard. You're digging into a field all day long for weeks and weeks and weeks, and you find something, you get super excited for 12 seconds. And at the 13 seconds, you're like, oh, yeah, that was obvious. And you go back to digging. I'm not a trained, like formally trained researcher, and it wasn't kind of a necessarily an ambition of me of creating, of having a research career. And I felt the hardness of it. I enjoyed a lot of like that a ton. But at the time, I decided that I wanted to go back to something more productive. And the other fun motivation was like, I mean, if we believe in AGI and if we believe the timelines might not be too long, it's actually the last train leaving the station to start a company. After that, it's going to be computers all the way down. And so that was kind of the true motivation for like trying to go there. So that's kind of the core motivation at the beginning of personally. And the motivation for starting a company was pretty simple. I had seen GPT-4 internally at the time, it was September 2022. So it was pre-GPT, but GPT-4 was ready since, I mean, I'd been ready for a few months internally. I was like, okay, that's obvious, the capabilities are there to create an insane amount of value to the world. And yet the deployment is not there yet. The revenue of OpenAI at the time were ridiculously small compared to what it is today. So the thesis was, there's probably a lot to be done at the product level to unlock the usage.Alessio [00:19:49]: Yeah. Let's talk a bit more about the form factor, maybe. I think one of the first successes you had was kind of like the WebGPT-like thing, like using the models to traverse the web and like summarize things. And the browser was really the interface. Why did you start with the browser? Like what was it important? And then you built XP1, which was kind of like the browser extension.Stan [00:20:09]: So the starting point at the time was, if you wanted to talk about LLMs, it was still a rather small community, a community of mostly researchers and to some extent, very early adopters, very early engineers. It was almost inconceivable to just build a product and go sell it to the enterprise, though at the time there was a few companies doing that. The one on marketing, I don't remember its name, Jasper. But so the natural first intention, the first, first, first intention was to go to the developers and try to create tooling for them to create product on top of those models. And so that's what Dust was originally. It was quite different than Lanchain, and Lanchain just beat the s**t out of us, which is great. It's a choice.Swyx [00:20:53]: You were cloud, in closed source. They were open source.Stan [00:20:56]: Yeah. So technically we were open source and we still are open source, but I think that doesn't really matter. I had the strong belief from my research time that you cannot create an LLM-based workflow on just one example. Basically, if you just have one example, you overfit. So as you develop your interaction, your orchestration around the LLM, you need a dozen examples. Obviously, if you're running a dozen examples on a multi-step workflow, you start paralyzing stuff. And if you do that in the console, you just have like a messy stream of tokens going out and it's very hard to observe what's going there. And so the idea was to go with an UI so that you could kind of introspect easily the output of each interaction with the model and dig into there through an UI, which is-Swyx [00:21:42]: Was that open source? I actually didn't come across it.Stan [00:21:44]: Oh yeah, it wasn't. I mean, Dust is entirely open source even today. We're not going for an open source-Swyx [00:21:48]: If it matters, I didn't know that.Stan [00:21:49]: No, no, no, no, no. The reason why is because we're not open source because we're not doing an open source strategy. It's not an open source go-to-market at all. We're open source because we can and it's fun.Swyx [00:21:59]: Open source is marketing. You have all the downsides of open source, which is like people can clone you.Stan [00:22:03]: But I think that downside is a big fallacy. Okay. Yes, anybody can clone Dust today, but the value of Dust is not the current state. The value of Dust is the number of eyeballs and hands of developers that are creating to it in the future. And so yes, anybody can clone it today, but that wouldn't change anything. There is some value in being open source. In a discussion with the security team, you can be extremely transparent and just show the code. When you have discussion with users and there's a bug or a feature missing, you can just point to the issue, show the pull request, show the, show the, exactly, oh, PR welcome. That doesn't happen that much, but you can show the progress if the person that you're chatting with is a little bit technical, they really enjoy seeing the pull request advancing and seeing all the way to deploy. And then the downsides are mostly around security. You never want to do security by obfuscation. But the truth is that your vector of attack is facilitated by you being open source. But at the same time, it's a good thing because if you're doing anything like a bug bountying or stuff like that, you just give much more tools to the bug bountiers so that their output is much better. So there's many, many, many trade-offs. I don't believe in the value of the code base per se. I think it's really the people that are on the code base that have the value and go to market and the product and all of those things that are around the code base. Obviously, that's not true for every code base. If you're working on a very secret kernel to accelerate the inference of LLMs, I would buy that you don't want to be open source. But for product stuff, I really think there's very little risk. Yeah.Alessio [00:23:39]: I signed up for XP1, I was looking, January 2023. I think at the time you were on DaVinci 003. Given that you had seen GPD 4, how did you feel having to push a product out that was using this model that was so inferior? And you're like, please, just use it today. I promise it's going to get better. Just overall, as a founder, how do you build something that maybe doesn't quite work with the model today, but you're just expecting the new model to be better?Stan [00:24:03]: Yeah, so actually, XP1 was even on a smaller one that was the post-GDPT release, small version, so it was... Ada, Babbage... No, no, no, not that far away. But it was the small version of GDPT, basically. I don't remember its name. Yes, you have a frustration there. But at the same time, I think XP1 was designed, was an experiment, but was designed as a way to be useful at the current capability of the model. If you just want to extract data from a LinkedIn page, that model was just fine. If you want to summarize an article on a newspaper, that model was just fine. And so it was really a question of trying to find a product that works with the current capability, knowing that you will always have tailwinds as models get better and faster and cheaper. So that was kind of a... There's a bit of a frustration because you know what's out there and you know that you don't have access to it yet. It's also interesting to try to find a product that works with the current capability.Alessio [00:24:55]: And we highlighted XP1 in our anatomy of autonomy post in April of last year, which was, you know, where are all the agents, right? So now we spent 30 minutes getting to what you're building now. So you basically had a developer framework, then you had a browser extension, then you had all these things, and then you kind of got to where Dust is today. So maybe just give people an overview of what Dust is today and the courtesies behind it. Yeah, of course.Stan [00:25:20]: So Dust, we really want to build the infrastructure so that companies can deploy agents within their teams. We are horizontal by nature because we strongly believe in the emergence of use cases from the people having access to creating an agent that don't need to be developers. They have to be thinkers. They have to be curious. But anybody can create an agent that will solve an operational thing that they're doing in their day-to-day job. And to make those agents useful, there's two focus, which is interesting. The first one is an infrastructure focus. You have to build the pipes so that the agent has access to the data. You have to build the pipes such that the agents can take action, can access the web, et cetera. So that's really an infrastructure play. Maintaining connections to Notion, Slack, GitHub, all of them is a lot of work. It is boring work, boring infrastructure work, but that's something that we know is extremely valuable in the same way that Stripe is extremely valuable because it maintains the pipes. And we have that dual focus because we're also building the product for people to use it. And there it's fascinating because everything started from the conversational interface, obviously, which is a great starting point. But we're only scratching the surface, right? I think we are at the pong level of LLM productization. And we haven't invented the C3. We haven't invented Counter-Strike. We haven't invented Cyberpunk 2077. So this is really our mission is to really create the product that lets people equip themselves to just get away all the work that can be automated or assisted by LLMs.Alessio [00:26:57]: And can you just comment on different takes that people had? So maybe the most open is like auto-GPT. It's just kind of like just trying to do anything. It's like it's all magic. There's no way for you to do anything. Then you had the ADAPT, you know, we had David on the podcast. They're very like super hands-on with each individual customer to build super tailored. How do you decide where to draw the line between this is magic? This is exposed to you, especially in a market where most people don't know how to build with AI at all. So if you expect them to do the thing, they're probably not going to do it. Yeah, exactly.Stan [00:27:29]: So the auto-GPT approach obviously is extremely exciting, but we know that the agentic capability of models are not quite there yet. It just gets lost. So we're starting, we're starting where it works. Same with the XP one. And where it works is pretty simple. It's like simple workflows that involve a couple tools where you don't even need to have the model decide which tools it's used in the sense of you just want people to put it in the instructions. It's like take that page, do that search, pick up that document, do the work that I want in the format I want, and give me the results. There's no smartness there, right? In terms of orchestrating the tools, it's mostly using English for people to program a workflow where you don't have the constraint of having compatible API between the two.Swyx [00:28:17]: That kind of personal automation, would you say it's kind of like an LLM Zapier type ofStan [00:28:22]: thing?Swyx [00:28:22]: Like if this, then that, and then, you know, do this, then this. You're programming with English?Stan [00:28:28]: So you're programming with English. So you're just saying, oh, do this and then that. You can even create some form of APIs. You say, when I give you the command X, do this. When I give you the command Y, do this. And you describe the workflow. But you don't have to create boxes and create the workflow explicitly. It just needs to describe what are the tasks supposed to be and make the tool available to the agent. The tool can be a semantic search. The tool can be querying into a structured database. The tool can be searching on the web. And obviously, the interesting tools that we're only starting to scratch are actually creating external actions like reimbursing something on Stripe, sending an email, clicking on a button in the admin or something like that.Swyx [00:29:11]: Do you maintain all these integrations?Stan [00:29:13]: Today, we maintain most of the integrations. We do always have an escape hatch for people to kind of custom integrate. But the reality is that the reality of the market today is that people just want it to work, right? And so it's mostly us maintaining the integration. As an example, a very good source of information that is tricky to productize is Salesforce. Because Salesforce is basically a database and a UI. And they do the f**k they want with it. And so every company has different models and stuff like that. So right now, we don't support it natively. And the type of support or real native support will be slightly more complex than just osing into it, like is the case with Slack as an example. Because it's probably going to be, oh, you want to connect your Salesforce to us? Give us the SQL. That's the Salesforce QL language. Give us the queries you want us to run on it and inject in the context of dust. So that's interesting how not only integrations are cool, and some of them require a bit of work on the user. And for some of them that are really valuable to our users, but we don't support yet, they can just build them internally and push the data to us.Swyx [00:30:18]: I think I understand the Salesforce thing. But let me just clarify, are you using browser automation because there's no API for something?Stan [00:30:24]: No, no, no, no. In that case, so we do have browser automation for all the use cases and apply the public web. But for most of the integration with the internal system of the company, it really runs through API.Swyx [00:30:35]: Haven't you felt the pull to RPA, browser automation, that kind of stuff?Stan [00:30:39]: I mean, what I've been saying for a long time, maybe I'm wrong, is that if the future is that you're going to stand in front of a computer and looking at an agent clicking on stuff, then I'll hit my computer. And my computer is a big Lenovo. It's black. Doesn't sound good at all compared to a Mac. And if the APIs are there, we should use them. There is going to be a long tail of stuff that don't have APIs, but as the world is moving forward, that's disappearing. So the core API value in the past has really been, oh, this old 90s product doesn't have an API. So I need to use the UI to automate. I think for most of the ICP companies, the companies that ICP for us, the scale ups that are between 500 and 5,000 people, tech companies, most of the SaaS they use have APIs. Now there's an interesting question for the open web, because there are stuff that you want to do that involve websites that don't necessarily have APIs. And the current state of web integration from, which is us and OpenAI and Anthropic, I don't even know if they have web navigation, but I don't think so. The current state of affair is really, really broken because you have what? You have basically search and headless browsing. But headless browsing, I think everybody's doing basically body.innertext and fill that into the model, right?Swyx [00:31:56]: MARK MIRCHANDANI There's parsers into Markdown and stuff.Stan [00:31:58]: FRANCESC CAMPOY I'm super excited by the companies that are exploring the capability of rendering a web page into a way that is compatible for a model, being able to maintain the selector. So that's basically the place where to click in the page through that process, expose the actions to the model, have the model select an action in a way that is compatible with model, which is not a big page of a full DOM that is very noisy, and then being able to decompress that back to the original page and take the action. And that's something that is really exciting and that will kind of change the level of things that agents can do on the web. That I feel exciting, but I also feel that the bulk of the useful stuff that you can do within the company can be done through API. The data can be retrieved by API. The actions can be taken through API.Swyx [00:32:44]: For listeners, I'll note that you're basically completely disagreeing with David Wan. FRANCESC CAMPOY Exactly, exactly. I've seen it since it's summer. ADEPT is where it is, and Dust is where it is. So Dust is still standing.Alessio [00:32:55]: Can we just quickly comment on function calling? You mentioned you don't need the models to be that smart to actually pick the tools. Have you seen the models not be good enough? Or is it just like, you just don't want to put the complexity in there? Like, is there any room for improvement left in function calling? Or do you feel you usually consistently get always the right response, the right parametersStan [00:33:13]: and all of that?Alessio [00:33:13]: FRANCESC CAMPOY So that's a tricky product question.Stan [00:33:15]: Because if the instructions are good and precise, then you don't have any issue, because it's scripted for you. And the model will just look at the scripts and just follow and say, oh, he's probably talking about that action, and I'm going to use it. And the parameters are kind of abused from the state of the conversation. I'll just go with it. If you provide a very high level, kind of an auto-GPT-esque level in the instructions and provide 16 different tools to your model, yes, we're seeing the models in that state making mistakes. And there is obviously some progress can be made on the capabilities. But the interesting part is that there is already so much work that can assist, augment, accelerate by just going with pretty simply scripted for actions agents. What I'm excited about by pushing our users to create rather simple agents is that once you have those working really well, you can create meta agents that use the agents as actions. And all of a sudden, you can kind of have a hierarchy of responsibility that will probably get you almost to the point of the auto-GPT value. It requires the construction of intermediary artifacts, but you're probably going to be able to achieve something great. I'll give you some example. We have our incidents are shared in Slack in a specific channel, or shipped are shared in Slack. We have a weekly meeting where we have a table about incidents and shipped stuff. We're not writing that weekly meeting table anymore. We have an assistant that just go find the right data on Slack and create the table for us. And that assistant works perfectly. It's trivially simple, right? Take one week of data from that channel and just create the table. And then we have in that weekly meeting, obviously some graphs and reporting about our financials and our progress and our ARR. And we've created assistants to generate those graphs directly. And those assistants works great. By creating those assistants that cover those small parts of that weekly meeting, slowly we're getting to in a world where we'll have a weekly meeting assistance. We'll just call it. You don't need to prompt it. You don't need to say anything. It's going to run those different assistants and get that notion page just ready. And by doing that, if you get there, and that's an objective for us to us using Dust, get there, you're saving an hour of company time every time you run it. Yeah.Alessio [00:35:28]: That's my pet topic of NPM for agents. How do you build dependency graphs of agents? And how do you share them? Because why do I have to rebuild some of the smaller levels of what you built already?Swyx [00:35:40]: I have a quick follow-up question on agents managing other agents. It's a topic of a lot of research, both from Microsoft and even in startups. What you've discovered best practice for, let's say like a manager agent controlling a bunch of small agents. It's two-way communication. I don't know if there should be a protocol format.Stan [00:35:59]: To be completely honest, the state we are at right now is creating the simple agents. So we haven't even explored yet the meta agents. We know it's there. We know it's going to be valuable. We know it's going to be awesome. But we're starting there because it's the simplest place to start. And it's also what the market understands. If you go to a company, random SaaS B2B company, not necessarily specialized in AI, and you take an operational team and you tell them, build some tooling for yourself, they'll understand the small agents. If you tell them, build AutoGP, they'll be like, Auto what?Swyx [00:36:31]: And I noticed that in your language, you're very much focused on non-technical users. You don't really mention API here. You mention instruction instead of system prompt, right? That's very conscious.Stan [00:36:41]: Yeah, it's very conscious. It's a mark of our designer, Ed, who kind of pushed us to create a friendly product. I was knee-deep into AI when I started, obviously. And my co-founder, Gabriel, was a Stripe as well. We started a company together that got acquired by Stripe 15 years ago. It was at Alain, a healthcare company in Paris. After that, it was a little bit less so knee-deep in AI, but really focused on product. And I didn't realize how important it is to make that technology not scary to end users. It didn't feel scary to me, but it was really seen by Ed, our designer, that it was feeling scary to the users. And so we were very proactive and very deliberate about creating a brand that feels not too scary and creating a wording and a language, as you say, that really tried to communicate the fact that it's going to be fine. It's going to be easy. You're going to make it.Alessio [00:37:34]: And another big point that David had about ADAPT is we need to build an environment for the agents to act. And then if you have the environment, you can simulate what they do. How's that different when you're interacting with APIs and you're kind of touching systems that you cannot really simulate? If you call it the Salesforce API, you're just calling it.Stan [00:37:52]: So I think that goes back to the DNA of the companies that are very different. ADAPT, I think, was a product company with a very strong research DNA, and they were still doing research. One of their goals was building a model. And that's why they raised a large amount of money, et cetera. We are 100% deliberately a product company. We don't do research. We don't train models. We don't even run GPUs. We're using the models that exist, and we try to push the product boundary as far as possible with the existing models. So that creates an issue. Indeed, so to answer your question, when you're interacting in the real world, well, you cannot simulate, so you cannot improve the models. Even improving your instructions is complicated for a builder. The hope is that you can use models to evaluate the conversations so that you can get at least feedback and you could get contradictive information about the performance of the assistance. But if you take actual trace of interaction of humans with those agents, it is even for us humans extremely hard to decide whether it was a productive interaction or a really bad interaction. You don't know why the person left. You don't know if they left happy or not. So being extremely, extremely, extremely pragmatic here, it becomes a product issue. We have to build a product that identifies the end users to provide feedback so that as a first step, the person that is building the agent can iterate on it. As a second step, maybe later when we start training model and post-training, et cetera, we can optimize around that for each of those companies. Yeah.Alessio [00:39:17]: Do you see in the future products offering kind of like a simulation environment, the same way all SaaS now kind of offers APIs to build programmatically? Like in cybersecurity, there are a lot of companies working on building simulative environments so that then you can use agents like Red Team, but I haven't really seen that.Stan [00:39:34]: Yeah, no, me neither. That's a super interesting question. I think it's really going to depend on how much, because you need to simulate to generate data, you need to train data to train models. And the question at the end is, are we going to be training models or are we just going to be using frontier models as they are? On that question, I don't have a strong opinion. It might be the case that we'll be training models because in all of those AI first products, the model is so close to the product surface that as you get big and you want to really own your product, you're going to have to own the model as well. Owning the model doesn't mean doing the pre-training, that would be crazy. But at least having an internal post-training realignment loop, it makes a lot of sense. And so if we see many companies going towards that all the time, then there might be incentives for the SaaS's of the world to provide assistance in getting there. But at the same time, there's a tension because those SaaS, they don't want to be interacted by agents, they want the human to click on the button. Yeah, they got to sell seats. Exactly.Swyx [00:40:41]: Just a quick question on models. I'm sure you've used many, probably not just OpenAI. Would you characterize some models as better than others? Do you use any open source models? What have been the trends in models over the last two years?Stan [00:40:53]: We've seen over the past two years kind of a bit of a race in between models. And at times, it's the OpenAI model that is the best. At times, it's the Anthropic models that is the best. Our take on that is that we are agnostic and we let our users pick their model. Oh, they choose? Yeah, so when you create an assistant or an agent, you can just say, oh, I'm going to run it on GP4, GP4 Turbo, or...Swyx [00:41:16]: Don't you think for the non-technical user, that is actually an abstraction that you should take away from them?Stan [00:41:20]: We have a sane default. So we move the default to the latest model that is cool. And we have a sane default, and it's actually not very visible. In our flow to create an agent, you would have to go in advance and go pick your model. So this is something that the technical person will care about. But that's something that obviously is a bit too complicated for the...Swyx [00:41:40]: And do you care most about function calling or instruction following or something else?Stan [00:41:44]: I think we care most for function calling because you want to... There's nothing worse than a function call, including incorrect parameters or being a bit off because it just drives the whole interaction off.Swyx [00:41:56]: Yeah, so got the Berkeley function calling.Stan [00:42:00]: These days, it's funny how the comparison between GP4O and GP4 Turbo is still up in the air on function calling. I personally don't have proof, but I know many people, and I'm probably part of them, to think that GP4 Turbo is still better than GP4O on function calling. Wow. We'll see what comes out of the O1 class if it ever gets function calling. And Cloud 3.5 Summit is great as well. They kind of innovated in an interesting way, which was never quite publicized. But it's that they have that kind of chain of thought step whenever you use a Cloud model or Summit model with function calling. That chain of thought step doesn't exist when you just interact with it just for answering questions. But when you use function calling, you get that step, and it really helps getting better function calling.Swyx [00:42:43]: Yeah, we actually just recorded a podcast with the Berkeley team that runs that leaderboard this week. So they just released V3.Stan [00:42:49]: Yeah.Swyx [00:42:49]: It was V1 like two months ago, and then they V2, V3. Turbo is on top.Stan [00:42:53]: Turbo is on top. Turbo is over 4.0.Swyx [00:42:54]: And then the third place is XLAM from Salesforce, which is a large action model they've been trying to popularize.Stan [00:43:01]: Yep.Swyx [00:43:01]: O1 Mini is actually on here, I think. O1 Mini is number 11.Stan [00:43:05]: But arguably, O1 Mini has been in a line for that. Yeah.Alessio [00:43:09]: Do you use leaderboards? Do you have your own evals? I mean, this is kind of intuitive, right? Like using the older model is better. I think most people just upgrade. Yeah. What's the eval process like?Stan [00:43:19]: It's funny because I've been doing research for three years, and we have bigger stuff to cook. When you're deploying in a company, one thing where we really spike is that when we manage to activate the company, we have a crazy penetration. The highest penetration we have is 88% daily active users within the entire employee of the company. The kind of average penetration and activation we have in our current enterprise customers is something like more like 60% to 70% weekly active. So we basically have the entire company interacting with us. And when you're there, there is so many stuff that matters most than getting evals, getting the best model. Because there is so many places where you can create products or do stuff that will give you the 80% with the work you do. Whereas deciding if it's GPT-4 or GPT-4 Turbo or et cetera, you know, it'll just give you the 5% improvement. But the reality is that you want to focus on the places where you can really change the direction or change the interaction more drastically. But that's something that we'll have to do eventually because we still want to be serious people.Swyx [00:44:24]: It's funny because in some ways, the model labs are competing for you, right? You don't have to do any effort. You just switch model and then it'll grow. What are you really limited by? Is it additional sources?Stan [00:44:36]: It's not models, right?Swyx [00:44:37]: You're not really limited by quality of model.Stan [00:44:40]: Right now, we are limited by the infrastructure part, which is the ability to connect easily for users to all the data they need to do the job they want to do.Swyx [00:44:51]: Because you maintain all your own stuff.Stan [00:44:53]: You know, there are companies out thereSwyx [00:44:54]: that are starting to provide integrations as a service, right? I used to work in an integrations company. Yeah, I know.Stan [00:44:59]: It's just that there is some intricacies about how you chunk stuff and how you process information from one platform to the other. If you look at the end of the spectrum, you could think of, you could say, oh, I'm going to support AirByte and AirByte has- I used to work at AirByte.Swyx [00:45:12]: Oh, really?Stan [00:45:13]: That makes sense.Swyx [00:45:14]: They're the French founders as well.Stan [00:45:15]: I know Jean very well. I'm seeing him today. And the reality is that if you look at Notion, AirByte does the job of taking Notion and putting it in a structured way. But that's the way it is not really usable to actually make it available to models in a useful way. Because you get all the blocks, details, et cetera, which is useful for many use cases.Swyx [00:45:35]: It's also for data scientists and not for AI.Stan [00:45:38]: The reality of Notion is that sometimes you have a- so when you have a page, there's a lot of structure in it and you want to capture the structure and chunk the information in a way that respects that structure. In Notion, you have databases. Sometimes those databases are real tabular data. Sometimes those databases are full of text. You want to get the distinction and understand that this database should be considered like text information, whereas this other one is actually quantitative information. And to really get a very high quality interaction with that piece of information, I haven't found a solution that will work without us owning the connection end-to-end.Swyx [00:46:15]: That's why I don't invest in, there's Composio, there's All Hands from Graham Newbig. There's all these other companies that are like, we will do the integrations for you. You just, we have the open source community. We'll do off the shelf. But then you are so specific in your needs that you want to own it.Swyx [00:46:28]: Yeah, exactly.Stan [00:46:29]: You can talk to Michel about that.Swyx [00:46:30]: You know, he wants to put the AI in there, but you know. Yeah, I will. I will.Stan [00:46:35]: Cool. What are we missing?Alessio [00:46:36]: You know, what are like the things that are like sneakily hard that you're tackling that maybe people don't even realize they're like really hard?Stan [00:46:43]: The real parts as we kind of touch base throughout the conversation is really building the infra that works for those agents because it's a tenuous walk. It's an evergreen piece of work because you always have an extra integration that will be useful to a non-negligible set of your users. I'm super excited about is that there's so many interactions that shouldn't be conversational interactions and that could be very useful. Basically, know that we have the firehose of information of those companies and there's not going to be that many companies that capture the firehose of information. When you have the firehose of information, you can do a ton of stuff with models that are just not accelerating people, but giving them superhuman capability, even with the current model capability because you can just sift through much more information. An example is documentation repair. If I have the firehose of Slack messages and new Notion pages, if somebody says, I own that page, I want to be updated when there is a piece of information that should update that page, this is not possible. You get an email saying, oh, look at that Slack message. It says the opposite of what you have in that paragraph. Maybe you want to update or just ping that person. I think there is a lot to be explored on the product layer in terms of what it means to interact productively with those models. And that's a problem that's extremely hard and extremely exciting.Swyx [00:48:00]: One thing you keep mentioning about infra work, obviously, Dust is building that infra and serving that in a very consumer-friendly way. You always talk about infra being additional sources, additional connectors. That is very important. But I'm also interested in the vertical infra. There is an orchestrator underlying all these things where you're doing asynchronous work. For example, the simplest one is a cron job. You just schedule things. But also, for if this and that, you have to wait for something to be executed and proceed to the next task. I used to work on an orchestrator as well, Temporal.Stan [00:48:31]: We used Temporal. Oh, you used Temporal? Yeah. Oh, how was the experience?Swyx [00:48:34]: I need the NPS.Stan [00:48:36]: We're doing a self-discovery call now.Swyx [00:48:39]: But you can also complain to me because I don't work there anymore.Stan [00:48:42]: No, we love Temporal. There's some edges that are a bit rough, surprisingly rough. And you would say, why is it so complicated?Swyx [00:48:49]: It's always versioning.Stan [00:48:50]: Yeah, stuff like that. But we really love it. And we use it for exactly what you said, like managing the entire set of stuff that needs to happen so that in semi-real time, we get all the updates from Slack or Notion or GitHub into the system. And whenever we see that piece of information goes through, maybe trigger workflows to run agents because they need to provide alerts to users and stuff like that. And Temporal is great. Love it.Swyx [00:49:17]: You haven't evaluated others. You don't want to build your own. You're happy with...Stan [00:49:21]: Oh, no, we're not in the business of replacing Temporal. And Temporal is so... I mean, it is or any other competitive product. They're very general. If it's there, there's an interesting theory about buy versus build. I think in that case, when you're a high-growth company, your buy-build trade-off is very much on the side of buy. Because if you have the capability, you're just going to be saving time, you can focus on your core competency, etc. And it's funny because we're seeing, we're starting to see the post-high-growth company, post-SKF company, going back on that trade-off, interestingly. So that's the cloud news about removing Zendesk and Salesforce. Do you believe that, by the way?Alessio [00:49:56]: Yeah, I did a podcast with them.Stan [00:49:58]: Oh, yeah?Alessio [00:49:58]: It's true.Swyx [00:49:59]: No, no, I know.Stan [00:50:00]: Of course they say it's true,Swyx [00:50:00]: but also how well is it going to go?Stan [00:50:02]: So I'm not talking about deflecting the customer traffic. I'm talking about building AI on top of Salesforce and Zendesk, basically, if I understand correctly. And all of a sudden, your product surface becomes much smaller because you're interacting with an AI system that will take some actions. And so all of a sudden, you don't need the product layer anymore. And you realize that, oh, those things are just databases that I pay a hundred times the price, right? Because you're a post-SKF company and you have tech capabilities, you are incentivized to reduce your costs and you have the capability to do so. And then it makes sense to just scratch the SaaS away. So it's interesting that we might see kind of a bad time for SaaS in post-hyper-growth tech companies. So it's still a big market, but it's not that big because if you're not a tech company, you don't have the capabilities to reduce that cost. If you're a high-growth company, always going to be buying because you go faster with that. But that's an interesting new space, new category of companies that might remove some SaaS. Yeah, Alessio's firmSwyx [00:51:02]: has an interesting thesis on the future of SaaS in AI.Alessio [00:51:05]: Service as a software, we call it. It's basically like, well, the most extreme is like, why is there any software at all? You know, ideally, it's all a labor interface where you're asking somebody to do something for you, whether that's a person, an AI agent or whatnot.Stan [00:51:17]: Yeah, yeah, that's interesting. I have to ask.Swyx [00:51:19]: Are you paying for Temporal Cloud or are you self-hosting?Stan [00:51:22]: Oh, no, no, we're paying, we're paying. Oh, okay, interesting.Swyx [00:51:24]: We're paying way too much.Stan [00:51:26]: It's crazy expensive, but it makes us-Swyx [00:51:28]: That's why as a shareholder, I like to hear that. It makes us go faster,Stan [00:51:31]: so we're happy to pay.Swyx [00:51:33]: Other things in the infrastack, I just want a list for other founders to think about. Ops, API gateway, evals, you know, anything interesting there that you build or buy?Stan [00:51:41]: I mean, there's always an interesting question. We've been building a lot around the interface between models and because Dust, the original version, was an orchestration platform and we basically provide a unified interface to every model providers.Swyx [00:51:56]: That's what I call gateway.Stan [00:51:57]: That we add because Dust was that and so we continued building upon and we own it. But that's an interesting question was in you, you want to build that or buy it?Swyx [00:52:06]: Yeah, I always say light LLM is the current open source consensus.Stan [00:52:09]: Exactly, yeah. There's an interesting question there.Swyx [00:52:12]: Ops, Datadog, just tracking.Stan [00:52:14]: Oh yeah, so Datadog is an obvious... What are the mistakes that I regret? I started as pure JavaScript, not TypeScript, and I think you want to, if you're wondering, oh, I want to go fast, I'll do a little bit of JavaScript. No, don't, just start with TypeScript. I see, okay.Swyx [00:52:30]: So interesting, you are a research engineer that came out of OpenAI that bet on TypeScript.Stan [00:52:36]: Well, the reality is that if you're building a product, you're going to be doing a lot of JavaScript, right? And Next, we're using Next as an example. It's

Screaming in the Cloud
Sleuthing Out the Key to Teamwork Dylan Etkin

Screaming in the Cloud

Play Episode Listen Later Nov 6, 2024 27:24


Corey Quinn chats with Dylan Etkin, CEO and co-founder of Sleuth. He joins this episode of Screaming Into the Cloud to share his insights on reshaping engineering metrics to prioritize team success. Sleuth emphasizes team-level productivity over individual output, sidestepping controversial metrics like lines of code and focusing on alignment and iterative improvement. By aggregating data from tools like GitHub, Jira, and Datadog, Sleuth provides actionable insights, helping leaders reallocate resources for optimal impact without disrupting unique team workflows. Designed for collaborative review, Sleuth's slide deck-like interface supports meaningful discussions around DORA metrics and deploy tracking. Show Highlights(0:00) Intro(0:51) Sleuth sponsor read(1:12) What Sleuth is(2:02) How Sleuth evaluates engineers' work(5:41) The value that evaluations brings to a business(9:34) Who Dylan usually discusses results with(11:04) Sleuth sponsor read(11:30) The day-to-day experience of using Sleuth(14:23) The importance of meeting people where they are(18:21) The actual outcome of implementing Sleuth(20:27) Why engineering teams should care about metrics(24:27) The interface that people have when they're working with Sleuth(26:23) Where you can find more from SleuthAbout Dylan EtkinDylan was one of the first twenty employees of Atlassian, and a founding engineer and the first architect of Jira. He has led engineering at scale for Bitbucket and Statuspage. He has a Master's in Computer Science from ASU. Dylan is a bit of a space nut and has been seen climbing around the inside of a life-size replica of the Mir space station in Star City Russia.SponsorSleuth: https://www.sleuth.io/ 

TechCrunch Startups – Spoken Edition
Datadog challenger Dash0 aims to dash observability bill shock

TechCrunch Startups – Spoken Edition

Play Episode Listen Later Nov 6, 2024 4:16


The end of zero-interest rates has driven companies to look for savings wherever they can, but one area continues to be a major budget drain. Observability — collecting and understanding data and systems — typically remains an organization's second-highest cloud expenditure, right after cloud provisioning itself. Learn more about your ad choices. Visit podcastchoices.com/adchoices

Capital
Empresas con identidad

Capital

Play Episode Listen Later Nov 4, 2024 17:58


Nos acompaña Cure51,una empresa de biotecnología que investiga los mecanismos biológicos responsables de la supervivencia en pacientes con cáncer de alto riesgo. Cure51 fue fundada en marzo de 2022 por Nicolas Wolikow y Simon Istolainen, junto con varios emprendedores como la doctora Paloma Cejas, experta en investigación medica en los hospitales de La Paz (Madrid) y Dana Farber (Boston), y cuatro centros de oncología de renombre mundial: Vall d'Hebron Instituto de Oncología (Barcelona - España), dos mas uno en París y otro en Lyon y otro en Berlín – Alemania.... Este año han obtenido una financiación semilla de 15 millones de euros que le permitirá , analizar e identificar los procesos moleculares que explican cómo unos pacientes con cáncer sobreviven durante periodos de tiempo muy prolongados a pesar de tener formas muy agresivas de la enfermedad. Esta compañía emplea las técnicas moleculares más punteras y colabora estrechamente con destacados centros oncológicos de todo el mundo entre los que se encuentran en el Vall d'Hebron Instituto de Oncología (VHIO). La compañía investiga cómo aprovechar esta nueva base de conocimiento para la medicina de precisión y desarrollar nuevos fármacos. La operación ha sido liderada por Sofinnova Partners, empresa europea de capital riesgo líder en ciencias de la vida, y cuenta con la participación de varios inversores como Hitachi Ventures GmbH, Life Extension Ventures, el magnate francés Xavier Niel y Olivier Pomel, CEO y cofundador de Datadog. Paloma Cejas, Cofundadora y directora de Investigación Traslacional de Cure51.

Screaming in the Cloud
Replay - Chaos Engineering for Gremlins with Jason Yee

Screaming in the Cloud

Play Episode Listen Later Oct 31, 2024 31:22


On this Replay, we're revisiting our conversation with Jason Yee, Staff Technical Advocate at Datadog. At the time of this recording, he was the Director of Advocacy at Gremlin, an enterprise-grade chaos engineering platform. Join Corey and Jason as they talk about what Gremlin is and what a director of advocacy does, making chaos engineering more accessible for the masses, how it's hard to calculate ROI for developer advocates, how developer advocacy and DevRel changes from one company to the next, why developer advocates need to focus on meaningful connections, why you should start chaos engineering as a mental game, qualities to look for in good developer advocates, the Break Things On Purpose podcast, and more.Show Highlights(0:00) Intro(0:31) Blackblaze sponsor read(0:58) The role of a Director of Advocacy(3:34) DevRel and twisting job definitions(5:50) How DevRel confusion manifests into marketing(11:37) Being able to measure and define a team's success(13:42) Building respect and a community in tech(15:22) Effectively courting a community(18:02) The challenges of Jason's job(21:06) Planning for failure modes(22:30) Determining your value in tech(25:41) The growth of Gremlin(30:16) Where you can find more from JasonAbout Jason YeeJason Yee is Staff Technical Avdocate at Datadog, where he works to inspire developers and ops engineers with the power of metrics and monitoring. Previously, he was the community manager for DevOps & Performance at O'Reilly Media and a software engineer at MongoDB.LinksBreak Things On Purpose podcast: https://www.gremlin.com/podcast/Twitter: https://twitter.com/gitbisectOriginal episodehttps://www.lastweekinaws.com/podcast/screaming-in-the-cloud/chaos-engineering-for-gremlins-with-jason-yee/SponsorBackblaze: https://www.backblaze.com/

The Official SaaStr Podcast: SaaS | Founders | Investors
SaaStr 765: The Bar Has Gone Up. The Era of HyperFunctional SaaS is Here with SaaStr CEO and Founder Jason Lemkin

The Official SaaStr Podcast: SaaS | Founders | Investors

Play Episode Listen Later Oct 16, 2024 44:47


SaaStr 765: The Bar Has Gone Up. The Era of HyperFunctional SaaS is Here with SaaStr CEO and Founder Jason Lemkin In this comprehensive episode, SaaStr CEO and Founder Jason Lemkin explains why the era of hyper-functional SaaS is here, and how it's reshaping the landscape of SaaS companies. Key topics include the critical role of AI and automation in meeting rising customer expectations and the imperative for tech companies to develop multi-product strategies to sustain growth. We explore how companies like Slice, Datadog, and Monday are navigating market dynamics, balancing efficiency with innovation, and adapting to accelerated benchmarks despite claims of downturns in SaaS spending. Additionally, we delve into the evolving sales challenges, maintaining product market fit, and finding untapped opportunities amidst the surge of AI advancements. Tune in to uncover strategies and real-world examples for building a thriving platform in the modern SaaS industry. -------------------------------------------------------------------------------------------- SaaStr hosts the largest SaaS community events on the planet. Hey everybody - thanks to the 10,000 of you who came out to SaaStr Annual. We had a blast and big news -- we'll be back in MAY of 2025. That's right, the SaaStr Annual will be a bit earlier next year, May 13-15 2025. We'll still be back in the same venue, in the SF bay area at the 40+ acre sprawling san mateo county events center. Grab your tickets at saastrannual.com with code jason50 for an extra discount on our very best pricing. --------------------------------------------------------------------------------------------  This episode is sponsored by: remote.com When the right person for the job is a world away, Remote Talent brings the world to you. As the top job board for remote-first companies, we give you powerful tools to post your listings and reach the world's top candidates and remote professionals. Start building your dream team from anywhere—visit Remote.com/jobs today. --------------------------------------------------------------------------------------------  This episode is sponsored by: Anrok A question for SaaS finance leaders, do you know where your customers are? Anrok tracks where your sales are creating exposure, and automates tax calculation and filing worldwide. Built for high-growth software companies, Anrok protects your revenue and saves you time. Visit anrok.com/saastr to learn more.

The Official SaaStr Podcast: SaaS | Founders | Investors
SaaStr 761: How to Build Pipeline and GTM Alignment with Datadog's CMO Sara Varni

The Official SaaStr Podcast: SaaS | Founders | Investors

Play Episode Listen Later Oct 2, 2024 29:59


SaaStr 761: How to Build Pipeline and GTM Alignment with Datadog's CMO Sara Varni Sara Varni, Datadog‘s CMO, shares how to build pipeline and create alignment across sales and marketing. The CMO role at Datadog covers the usual aspects of marketing, including product marketing, corporate communications, events, partnership marketing, and solutions marketing. While these insights seem like pipeline 101, they aren't always implemented. Let's look at some quick facts about Sara's experience. She spent a decade at Salesforce before moving on to Twilio as CMO when they were at about $400M ARR with 30 sales reps. They had a healthy self-service motion, but growth was flattening. About four years later, she joined Attentive as CMO, a late-stage startup in the SMS marketing space. Now, she's the CMO at Datadog. Each of these companies had different GTM motions.   ----------------------------------------------------------------------------------------------- Hey everybody - thanks to the 10,000 of you who came out to SaaStr Annual. We had a blast and big news -- we'll be back in MAY of 2025. That's right, the SaaStr Annual will be a bit earlier next year, May 13-15 2025. We'll still be back in the same venue, in the SF bay area at the 40+ acre sprawling san mateo county events center. Grab your tickets at saastrannual.com with code jason50 for an extra discount on our very best pricing. --------------------------------------------------------------------------------------------  This episode is sponsored by: remote.com When the right person for the job is a world away, Remote Talent brings the world to you. As the top job board for remote-first companies, we give you powerful tools to post your listings and reach the world's top candidates and remote professionals. Start building your dream team from anywhere—visit Remote.com/jobs today. --------------------------------------------------------------------------------------------  This episode is sponsored by: Anrok A question for SaaS finance leaders, do you know where your customers are? Anrok tracks where your sales are creating exposure, and automates tax calculation and filing worldwide. Built for high-growth software companies, Anrok protects your revenue and saves you time. Visit anrok.com/saastr to learn more.

Sales Is King
188: The New Sales Landscape | Datadog's Frank Perkins

Sales Is King

Play Episode Listen Later Oct 1, 2024 67:26


In this conversation, Dan Sixsmith and(Chief Of Staff to) Datadog CRO Frank Perkins discuss the evolving landscape of sales, particularly in the tech industry. They explore the impact of COVID-19 on buyer behavior, the importance of emotional connections in sales, and the role of AI in enhancing sales processes. Frank shares insights from his career, including his experiences at Apple and Salesforce, and emphasizes the significance of personal branding for sellers. The discussion also touches on the challenges of measuring sales success and the importance of integrity and trust in sales relationships. Takeaways Sales today is influenced heavily by buyer behavior changes post-COVID. Emotional connection is crucial in the sales process. AI is set to revolutionize sales through opportunity scoring and prospecting. Sales leaders must adapt their methodologies to meet new buyer expectations. Personal branding is important for sellers to establish trust and credibility. The role of Chief of Staff in sales organizations is becoming increasingly vital. Understanding account intelligence can significantly improve sales strategies. Sales success is defined by the ability to create compelling events for buyers. Sales processes should be flexible and adaptable to changing market conditions. Integrity and trust are foundational to successful sales relationships. Chapters 00:00 Introduction and Current State of Sales 02:14 The Impact of COVID-19 on Sales and Buyer Behavior 03:40 The Changing Landscape of Sales and Buyer Behavior 08:31 The Role of Sales Leaders and Process 11:51 Building a Strong Sales Process and Methodology 18:55 The Importance of Account Intelligence 24:00 The Potential of AI in Sales 26:21 AI in Opportunity Scoring, Outbound Prospecting, and Account Intelligence 31:48 Creating a Compelling Event: Driving Action in Sales 35:08 Addressing Customer Pain Points 36:39 The Importance of Urgency and Commitment 37:33 Finding Legitimate Pain and Solutions 39:11 Prioritizing Pain over the Deal 40:49 The Role of Emotional Connection and Passion 50:34 The Power of Thoughtfulness in Sales 59:25 Building Trust and Integrity 01:05:15 Defining Success: Strong Connections and Trusted Guidance 01:07:08 lifestyle-outro-low.wav

Absolute AppSec
Episode 262 - w/ Ariel Shin - Building a Security Program

Absolute AppSec

Play Episode Listen Later Oct 1, 2024


Ariel Shin joins Ken Johnson (@cktricky on social media) and Seth Law (@sethlaw) for a special episode of Absolute AppSec. Ariel is currently a Security Engineering Manager at Datadog after a three-year stint at Twilio where she worked as an engineering manager in product security, a product security team lead, and a senior product security engineer. This year at Bsides SF 2024, she presented on her time at Twilio in a retrospective talk entitled “Six Years in Review: Transforming Company Culture to Embrace Risk.” The video from Bsides SF can be found here: https://www.youtube.com/watch?v=cQE1OqCpeI8. Before Twilio, Ariel worked at one medical as an appsec engineer as well as spending time as a Technology and Privacy consultant with Protiviti. She also helps build the professional appsec and prodsec communities as a frequent commenter and presenter at security conferences.

The MAD Podcast with Matt Turck
AI at Datadog: Monitoring machines in the age of LLMs | Olivier Pomel, CEO of Datadog

The MAD Podcast with Matt Turck

Play Episode Listen Later Sep 27, 2024 61:30


In this episode, we dive deep into the story of how Datadog evolved from a single product to a multi-billion dollar observability platform with its co-founder, Olivier Pomel. Olivier shares exclusive insights on Datadog's unique approach to product development—why they avoid the "Apple approach" of building in secret and instead work closely with customers from day one. You'll hear about the early days when Paul Graham of Y Combinator turned down Datadog, questioning their lack of a first product. Olivier also reveals the strategies behind their iterative product launches and why they insist on charging early to ensure they're delivering real value. The second half of the conversation is focused on all things AI and data at Datadog - the company's initial reluctance to use AI in its products, how Generative AI changed everything, and Datadog's current AI efforts including Watchdog, Bits AI and Toto, their new time series foundational model. We close the episode by asking Olivier about his thoughts on the topic du jour: founder mode! ▶️ Listen to 2020 Data Driven NYC episode with Oliver Pomel: https://www.youtube.com/watch?v=oXKEFHeEvMs DATADOG Website - https://www.datadoghq.com Twitter - https://x.com/datadoghq Olivier Pomel LinkedIn - https://www.linkedin.com/in/olivierpomel Twitter - https://x.com/oliveur FIRSTMARK Website - https://firstmark.com Twitter - https://twitter.com/FirstMarkCap Matt Turck (Managing Director) LinkedIn - https://www.linkedin.com/in/turck/ Twitter - https://twitter.com/mattturck

Real World Serverless with theburningmonk
#108: Lambda on Rust with James Eastham

Real World Serverless with theburningmonk

Play Episode Listen Later Sep 20, 2024 62:02


Thank you to Hookdeck for sponsoring this episode. If you're looking to level-up your event-driven architecture, then check out their serverless event gateway at hookdeck.com/theburningmonk and help support this channel.James Eastham is a developer advocate at Datadog and co-author of "Crafting Lambda Functions in Rust". In this episode, we dive into writing Lambda functions in Rust and why you should invest in learning Rust.Links from the episode:NSA whitepaper on memory safetyJulian Wood's Lambda internals talk at re:Invent 2022Jame's YouTube channelCrafting Lambda Functions in RustEp106 on middy with Luciano MamminoEp97 on LLRT (the superfast JavaScript runtime for Lambda)Opening theme song:Cheery Monday by Kevin MacLeodLink: https://incompetech.filmmusic.io/song/3495-cheery-mondayLicense: http://creativecommons.org/licenses/by/4.

The Twenty Minute VC: Venture Capital | Startup Funding | The Pitch
20VC: Index's Shardul Shah on Why Market Size is a Trap | Biggest Lessons on Pricing from Leading Rounds in Wiz & Datadog | Why Benchmarks & Averages in VC are BS | How Index Makes Decisions and Why Growth & Early are the Same Investing Style

The Twenty Minute VC: Venture Capital | Startup Funding | The Pitch

Play Episode Listen Later Sep 16, 2024 53:05


Shardul Shah is a Partner at Index Ventures and one of the greatest cyber security investors of the last two decades. Among his many wins, Shardul has led rounds in Datadog, Wiz, Duo Security, Coalition and more. Shardul is also the only Partner investing at Index to have worked in every single Index office from London, to SF, to NYC to Geneva. Prior to Index, Shardul worked with Summit Partners, focusing on healthcare and internet technologies. In Today's Episode with Shardul Shah We Discuss: 1. Investing Lessons from Wiz and Datadog: Why does Shardul believe that TAM (total addressable market) is BS? Why does Shardul believe that every great deal will be expensive? How does Shardul evaluate when to double down and concentrate capital vs when to let someone else come in and lead a round in an existing company? How does Shardul think about when is the right time to sell a position in a company? 2. How the Best VCs Make Decisions: How does Shardul and Index create an environment of truth-seeking together, that is optimised for the best decision-making to take place? What are the biggest mistakes in how VCs make decisions today? Why does Shardul believe that all first meetings should be 30 mins not 60 mins? Why does Shardul believe it is so much harder to make investment decisions when partnerships are remote? What is better remote? 3. The Core Pillars of Venture: Sourcing, Selecting, Securing and Servicing: Which one does Shardul believe he is best at? What is he worst at? Does Shardul believe with the downturn we have moved into a world of selection and not just winning every new deal? Does Shardul believe that VCs provide any value? What are the biggest misnomers when it comes to "VC value add"? 4. Lessons from the Best Investors in the World: Who is the best board member that Shardul sits on a board with? What has Shardul learned from Gili Raanan and Doug Leone on being a good board member? What have been some of Shardul's biggest investing lessons from Danny Rimer? Why does Shardul hate benchmarks when it comes to investing?  

PodRocket - A web development podcast from LogRocket
Production horror stories with Dan Neciu

PodRocket - A web development podcast from LogRocket

Play Episode Listen Later Aug 22, 2024 27:17


Dan Neciu, technical co-founder and tech lead of CareerOS, shares intriguing production horror stories, discusses the importance of rigorous testing, and provides valuable insights into preventing and managing software bugs in both backend and frontend development. Links https://neciudan.dev https://www.youtube.com/@NeciuDan https://www.linkedin.com/in/neciudan https://x.com/neciudan We want to hear from you! How did you find us? Did you see us on Twitter? In a newsletter? Or maybe we were recommended by a friend? Let us know by sending an email to our producer, Emily, at emily.kochanekketner@logrocket.com (mailto:emily.kochanekketner@logrocket.com), or tweet at us at PodRocketPod (https://twitter.com/PodRocketpod). Follow us. Get free stickers. Follow us on Apple Podcasts, fill out this form (https://podrocket.logrocket.com/get-podrocket-stickers), and we'll send you free PodRocket stickers! What does LogRocket do? LogRocket provides AI-first session replay and analytics that surfaces the UX and technical issues impacting user experiences. Start understand where your users are struggling by trying it for free at [LogRocket.com]. Try LogRocket for free today.(https://logrocket.com/signup/?pdr) Special Guest: Dan Neciu.

Software Defined Talk
Episode 480: No offsite content

Software Defined Talk

Play Episode Listen Later Aug 16, 2024 76:04


This week, we revisit our 2021 Intel CEO predictions, discuss Hyperscaler AI investment concerns and debate LinkedIn content. Plus, which products could thrive if Google were broken up? Watch the YouTube Live Recording of Episode (https://www.youtube.com/watch?v=q6rXxIXc1Us) 480 (https://www.youtube.com/watch?v=q6rXxIXc1Us) Runner-up Titles New camera, dog is still here Teenagers can put away some meat We operated at a loss last year because of BBQ Dog ate your Internet Getting out of a Zoom Under the couch data center Excuse me can I plug in my Satellite Dish A lot of 86-stuff You shouldn't be AI-ing that much Out of tokens, again Never bet against inertia Home of Markdown One cord, all power. Rundown Looking back at Intel Predicatons SDT Episode 281 (https://www.softwaredefinedtalk.com/281) Gaudi processors & Intel's AI portfolio (https://changelog.com/practicalai/281) How chip giant Intel spurned OpenAI and fell behind the times (https://finance.yahoo.com/news/chip-giant-intel-spurned-openai-110138000.html) Is AI too expensive for HyperScalers? Microsoft's AI Dreams Make for an Expensive Reality (https://www.wsj.com/tech/ai/microsofts-ai-dreams-make-for-an-expensive-reality-90a8c8e4) Amazon's cloud unit reports 19% revenue growth, topping estimates (https://www.cnbc.com/2024/08/01/aws-q2-earnings-report.html) Big Tech Fails to Convince Wall Street That AI Is Paying Off (https://www.bloomberg.com/news/articles/2024-08-02/big-tech-fails-to-convince-wall-street-that-ai-is-paying-off?srnd=phx-technology) Government shelves £1.3bn UK tech and AI plans (https://www.bbc.com/news/articles/cyx5x44vnyeo?_bhlid=027b7775b0cab914048391ea5d8c6254d1468ee1) Are all LinkedIn posts just a form of virtue signaling? There's One Thing You Really Don't Need to Do After Getting Laid Off (https://slate.com/technology/2024/07/laid-off-jobs-announcement-grateful-twitter-linkedin.html) Lighting round US Considers a Rare Antitrust Move: Breaking Up Google (https://www.bloomberg.com/news/articles/2024-08-13/doj-considers-seeking-google-goog-breakup-after-major-antitrust-win?srnd=homepage-americas) H-E-B is finally testing Apple Pay, (https://www.threads.net/@dsilverman/post/C-fVfKQJFyr/?xmt=AQGzlkdYqDg5ncWBmOvviB_QPKGOhv4CndLpOQtMpNlYLg) Relevant to your Interests Why CSV is still king (https://konbert.com/blog/why-csv-is-still-king?utm_source=changelog-news) Missing Semester IAP 2020 (https://www.youtube.com/playlist?list=PLyzOVJj3bHQuloKGG59rS43e29ro7I57J) iOS 18 has fulfilled my dream of destroying ads with a Thanos snap (https://www.theverge.com/2024/8/6/24214338/apple-ios-18-thanos-snap-animation-hide-ads) Sonos reportedly delaying two upcoming product releases to fix its misbehaving app (https://www.phonearena.com/news/sonos-reportedly-delaying-two-upcoming-product-releases-to-fix-its-misbehaving-app_id161241) Apple's Mac Mini With M4 Chip Will Be Its Smallest Computer Ever (https://www.bloomberg.com/news/articles/2024-08-08/mac-mini-m4-apple-plans-to-release-smallest-desktop-computer-yet) Background check company breached, nearly 3 billion exposed in data theft (https://mashable.com/article/background-check-company-breached-3-billion-affected) Apple's new USB-C charging has driven us all to Charging Cord Hell (https://apple.news/Awy0Ps7UvQSGZ8ASo2bhTtA) The Wizard of Cyber: what is behind Wiz's success (https://ventureinsecurity.net/p/the-wizard-of-cyber-what-is-behind) Brands should avoid this popular term. It's turning off customers (https://www.cnn.com/2024/08/10/business/brands-avoid-term-customers/index.html) Datadog, MongoDB, Snowflake in spotlight as cloud consumption rose in July: BofA (https://seekingalpha.com/news/4138587-datadog-mongodb-snowflake-in-spotlight-cloud-consumption-rose-in-july-bofa) The Elon / Trump interview on X started with an immediate tech disaster (https://www.theverge.com/2024/8/12/24219121/donald-trump-elon-musk-interview-x-twitter-crashes) Google takes another startup out of the AI race (https://www.theverge.com/2024/8/2/24212348/google-hires-character-ai-noam-shazeer) How Big Tech is swallowing the AI industry (https://www.theverge.com/2024/7/1/24190060/amazon-adept-ai-acquisition-playbook-microsoft-inflection) Delta CEO offers employees free flights after CrowdStrike-Microsoft chaos (https://www.cnbc.com/2024/08/02/delta-ceo-offers-employees-free-flights-after-crowdstrike-microsoft-chaos.html) Microsoft says Delta ignored Satya Nadella's offer of CrowdStrike help (https://www.theverge.com/2024/8/6/24214371/microsoft-delta-letter-crowdstrike-response-comments) External Technical Root Cause Analysis (https://www.crowdstrike.com/wp-content/uploads/2024/08/Channel-File-291-Incident-Root-Cause-Analysis-08.06.2024.pdf) Nonsense not sure why barking at it didn't help tbh. 12/10 for both #weratedogs (https://www.tiktok.com/t/ZTNgm76JG/) Starbucks replaces CEO Laxman Narasimhan with Chipotle CEO Brian Niccol (https://www.cnbc.com/2024/08/13/starbucks-replaces-ceo-laxman-narasimhan-with-chipotle-ceo-brian-niccol.html) Listener Feedback How South of the Border Keeps Going After All These Years (https://www.theassemblync.com/place/south-of-the-border/) Apple's new USB-C charging has driven us all to Charging Cord Hell (https://apple.news/Awy0Ps7UvQSGZ8ASo2bhTtA) XBox (https://www.xbox.com/en-US/xbox-game-pass/ultimate) (https://www.xbox.com/en-US/xbox-game-pass/ultimate)G (https://www.xbox.com/en-US/xbox-game-pass/ultimate)ame (https://www.xbox.com/en-US/xbox-game-pass/ultimate) P (https://www.xbox.com/en-US/xbox-game-pass/ultimate)ass (https://www.xbox.com/en-US/xbox-game-pass/ultimate) U (https://www.xbox.com/en-US/xbox-game-pass/ultimate)ltimate (https://www.xbox.com/en-US/xbox-game-pass/ultimate) Conferences DevOpsDays Birmingham (https://devopsdays.org/events/2024-birmingham-al/welcome/), Aug 19-21, 2024 SpringOne (https://springone.io/?utm_source=cote&utm_campaign=devrel&utm_medium=newsletter&utm_content=newsletterUpcoming)/VMware Explore US (https://blogs.vmware.com/explore/2024/04/23/want-to-attend-vmware-explore-convince-your-manager-with-these/?utm_source=cote&utm_campaign=devrel&utm_medium=newsletter&utm_content=newsletterUpcoming), Aug 26-29, 2024 DevOpsDays Antwerp (https://devopsdays.org/events/2024-antwerp/welcome/), Sept 4–5, 2024, 15th anniversary Cloud Foundry Day EU (https://events.linuxfoundation.org/cloud-foundry-day-europe/), Karlsruhe, GER, Oct 9, 2024, 20% off with code CFEU24VMW Civo Navigate Europe, Berlin (https://www.civo.com/navigate/europe), Sept 10-11, 2024 SREday London 2024 (https://sreday.com/2024-london/), Sept 19–20, 2024. Coté speaking, 20% off with the code SRE20DAY (https://sreday.com/2024-london/#tickets) SDT News & Community Join our Slack community (https://softwaredefinedtalk.slack.com/join/shared_invite/zt-1hn55iv5d-UTfN7mVX1D9D5ExRt3ZJYQ#/shared-invite/email) Email the show: questions@softwaredefinedtalk.com (mailto:questions@softwaredefinedtalk.com) Free stickers: Email your address to stickers@softwaredefinedtalk.com (mailto:stickers@softwaredefinedtalk.com) Follow us on social media: Twitter (https://twitter.com/softwaredeftalk), Threads (https://www.threads.net/@softwaredefinedtalk), Mastodon (https://hachyderm.io/@softwaredefinedtalk), LinkedIn (https://www.linkedin.com/company/software-defined-talk/), BlueSky (https://bsky.app/profile/softwaredefinedtalk.com) Watch us on: Twitch (https://www.twitch.tv/sdtpodcast), YouTube (https://www.youtube.com/channel/UCi3OJPV6h9tp-hbsGBLGsDQ/featured), Instagram (https://www.instagram.com/softwaredefinedtalk/), TikTok (https://www.tiktok.com/@softwaredefinedtalk) Book offer: Use code SDT for $20 off "Digital WTF" by Coté (https://leanpub.com/digitalwtf/c/sdt) Sponsor the show (https://www.softwaredefinedtalk.com/ads): ads@softwaredefinedtalk.com (mailto:ads@softwaredefinedtalk.com) Recommendations Brandon: Google Colab (https://colab.google) Matt: macOS Universal Control (https://support.apple.com/en-us/102459) Coté: Present & Correct (https://maps.app.goo.gl/NXJNw4TgGdFqcJPq6?g_st=com.google.maps.preview.copy> -

Screaming in the Cloud
Summer Replay - That Datadog Will Hunt with Dann Berg

Screaming in the Cloud

Play Episode Listen Later Aug 15, 2024 37:31


In this Screaming in the Cloud Summer Replay, we revisit our conversation with Dann Berg. At the time, he was a Senior Cloud Analyst at Datadog, but he now provides community support for the FinOps Foundation. Dann and Corey go into the weeds of cost optimization, and each of them bring their respective experiences forward. Dann's offers his take on multi-cloud and how Datadog is tackling its customer needs there. But the talent doesn't end there, Dann is also an emerging thinker and influencer in the space, and to boot, an accomplished writer and playwright. Two of his plays have been produced in NYC and China. Check out their conversation!Show Highlights:(0:00) Intro(1:02) Duckbill Group sponsor read(1:36) Transitioning to Senior Cloud Ops Analyst(5:12) The composition of Dann's team(6:54) Cloud cost optimization in the regular business cycle(10:43) Helping customers understand their cloud bills(17:42) Paying attention to pricing changes(21:06) The psychology of cloud economics(23:20) Working with multiple clouds(25:02) Duckbill Group sponsor read(25:46) Spending too much money to save too little money(31:12) The dangers of relying on third-party tools(34:01) Pricing woes(36:25) Where you can find DannAbout Dann BergDann Berg currently works part-time with FinOps after spending more than a decade in the industry. He is also an active member of the larger technical community, hosting the monthly New York City FinOps Meetup, and has been published multiple times in places such as MSNBC, Fox News, NPR, and others. When he's not saving companies millions of dollars, he's writing plays, and has had two full-lengh plays produced in New York City and China., Dann is the Director of Community at Vantage. Previously, first FinOps Practitioner at Datadog and FullStory. Host of the NYC FinOps Meetup for almost three years. He also writes plays.Links:Datadog: https://www.datadoghq.comPersonal Website: https://dannb.orgLinkedIn: https://www.linkedin.com/in/dannberg/Twitter: https://twitter.com/dannbergMonthly newsletter: https://dannb.org/newsletter/Previous SITC episode with Dann Berg, Episode 51: https://www.lastweekinaws.com/podcast/screaming-in-the-cloud/episode-51-size-of-cloud-bill-not-about-number-of-customers-but-number-of-engineers-you-ve-hired/Original Episode:https://www.lastweekinaws.com/podcast/screaming-in-the-cloud/that-datadog-will-hunt-with-dann-berg/Sponsor:The Duckbill Group: https://www.duckbillgroup.com/

The Drill Down
Drill Down Earnings, Ep. 188: Datadog Q2 earnings essentials ($DDOG)

The Drill Down

Play Episode Listen Later Aug 8, 2024 6:17


Instant analysis of Datadog ($DDOG) Q2 earnings, as we hear from CEO Olivier Pomel.  More than “beat” or “miss” –the Drill Down Earnings with Futurum Group chief market strategist Cory Johnson has the business stories behind stocks on the move.    https://x.com/corytv #Datadog #Earnings @Datadog $DDOG #Technology #Software #CloudComputing #Chips #AI #ArtificialIntelligence #Semiconductors #Stocks #Trading #Business @DrillDownPod Learn more about your ad choices. Visit megaphone.fm/adchoices

Software Defined Talk
Episode 477: We're an N-1 Organization

Software Defined Talk

Play Episode Listen Later Jul 26, 2024 70:38


This week, we discuss the CrowdStrike outage, FinOps data exports, and the state of open-source forks. Plus, Matt shares some exciting exclusive news about his future! Watch the YouTube Live Recording of Episode (https://www.youtube.com/watch?v=hYoFk0K_XpI) 477 (https://www.youtube.com/watch?v=hYoFk0K_XpI) Runner-up Titles Matt Ray Explains Channel File 291 Documenting CYA An intern did it Default lifestyle strikes again All the Nelson GIFs Rundown CrowdStrike Huge Microsoft Outage Linked to CrowdStrike Takes Down Computers Around the World (https://www.wired.com/story/microsoft-windows-outage-crowdstrike-global-it-probems/) 12-hour timelapse of airline traffic after what was likely the biggest IT outage in history (https://x.com/US_Stormwatch/status/1814268813879206397) Flights grounded and offices hit as internet users face disruptions (https://apnews.com/live/internet-global-outage-crowdstrike-microsoft-downtime) TODAY (@TODAYshow) on X (https://x.com/TODAYshow/status/1814266372882391523?ref_src=twsrc%5Egoogle%7Ctwcamp%5Eserp%7Ctwgr%5Etweet) George Kurtz (@George_Kurtz) on X (https://x.com/George_Kurtz/status/1814316045185822981) CrowdStrike's Global Outage Doesn't Have to Be a Recurring Nightmare (https://www.bloomberg.com/opinion/articles/2024-07-19/crowdstrike-s-nightmare-it-microsoft-outage-shouldn-t-be-normal?srnd=homepage-americas) Heard on the Street: CrowdStrike May Get More Than a Slap (https://www.wsj.com/livecoverage/stock-market-today-dow-sp500-nasdaq-live-07-19-2024/card/heard-on-the-street-crowdstrike-may-get-more-than-a-slap-CbyAd5zi7ELT4miAZHNV) What Happened to Digital Resilience? (https://www.nytimes.com/2024/07/19/us/politics/crowdstrike-outage.html?unlocked_article_code=1.8k0._ZDj.e5unf_bqIJNo&smid=url-share) SolarWinds Defeats Part of SEC's Fraud Case Over Hack (https://www.wsj.com/articles/solarwinds-defeats-part-of-secs-fraud-case-over-hack-ec69169a) Technical Details: Falcon Update for Windows Hosts (https://www.crowdstrike.com/blog/falcon-update-for-windows-hosts-technical-details/) Microsoft tried to get AV vendors to use APIs (https://www.threads.net/@sbisson/post/C9pIIYmo19q?xmt=AQGzVYTNKy9-De3zRXlIsl7QNqarqWsTWlmD_4Wc-7MM2A) House committee calls on CrowdStrike CEO to testify on global outage (https://www.washingtonpost.com/technology/2024/07/22/house-committee-calls-crowdstrike-ceo-testify-global-outage/) Crashes and Competition (https://stratechery.com/2024/crashes-and-competition/) The CrowdStrike Failure Was a Warning (https://www.theatlantic.com/ideas/archive/2024/07/crowdstrike-failure-warning-solutions/679174/) Defective McAfee update causes worldwide meltdown of XP PCs (https://www.zdnet.com/article/defective-mcafee-update-causes-worldwide-meltdown-of-xp-pcs/?utm_source=newsletter&utm_medium=email&utm_campaign=newsletter_axioscodebook&stream=top) CrowdStrike broke Debian and Rocky Linux months ago, but no one noticed (https://www.neowin.net/news/crowdstrike-broke-debian-and-rocky-linux-months-ago-but-no-one-noticed/#google_vignette) CrowdStrike Update: Latest News, Lessons Learned from a Retired Microsoft Engineer (https://youtu.be/ZHrayP-Y71Q?si=AmavOuoU_IjGMTFi) CrowdStrike offers a $10 apology gift card to say sorry for outage (https://techcrunch.com/2024/07/24/crowdstrike-offers-a-10-apology-gift-card-to-say-sorry-for-outage/) Announcing Data Exports for FOCUS 1.0 (Preview) in AWS Billing and Cost Management (https://aws.amazon.com/blogs/aws-cloud-financial-management/announcing-data-exports-for-focus-1-0-preview-in-aws-billing-and-cost-management/) Wiz walks away from $23 billion deal with Google, will pursue IPO (https://www.cnbc.com/2024/07/23/google-wiz-deal-dead.html) Import and export Markdown in Google Docs (http://workspaceupdates.googleblog.com/2024/07/import-and-export-markdown-in-google-docs.html) Google URL Shortener links will no longer be available (https://developers.googleblog.com/en/google-url-shortener-links-will-no-longer-be-available/) The Post-Valkey World (https://redmonk.com/sogrady/2024/07/16/post-valkey-world/) A tale of two forks - comparing Valkey/Redis and OpenTofu/Terraform! (https://www.linkedin.com/posts/danlorenc_oss-opensource-community-activity-7221488717704609792-U2SR/?utm_source=share&utm_medium=member_desktop) Datadog rumoured to be sniffing round GitLab as tech M&A market heats up (https://www.thestack.technology/datadog-rumoured-to-be-sniffing-round-gitlab-as-tech-m-a-market-heats-up/) Google-Backed Software Developer GitLab Eyes Sale, Reuters Says (https://www.bloomberg.com/news/articles/2024-07-17/google-backed-software-developer-gitlab-eyes-sale-reuters-says) Relevant to your Interests Google Open Sources 27B Parameter Gemma 2 Language Model (https://www.infoq.com/news/2024/07/google-gemma-2/) What It Really Takes to Build an AI Datacenter (https://www.bloomberg.com/news/articles/2024-06-21/what-it-really-takes-to-build-an-ai-datacenter) State of Developer Experience 2024 (https://newsletter.getdx.com/p/state-of-developer-experience-2024?r=2d4o&utm_campaign=post&utm_medium=web) The Return-to-Office Productivity Argument Is Over (https://www.inc.com/joe-procopio/the-return-to-office-productivity-argument-is-over.html) A new path for Privacy Sandbox on the web (https://privacysandbox.com/news/privacy-sandbox-update) The search for the random numbers that run our lives (https://www.bbc.com/future/article/20240704-the-search-for-the-random-numbers-that-run-our-lives) OpenAI is releasing a cheaper, smarter model (https://www.theverge.com/2024/7/18/24200714/openai-new-cheaper-smarter-model-gpt-4o-mini) Microsoft unveils a large language model that excels at encoding spreadsheets (https://www.thestack.technology/microsoft-llm-spreadsheet-llm/) Maestro: Netflix's Workflow Orchestrator (https://netflixtechblog.com/maestro-netflixs-workflow-orchestrator-ee13a06f9c78) IBM shares jump on earnings and revenue beat (https://www.cnbc.com/2024/07/24/ibm-q2-earnings-report-2024.html) US banks to begin reporting Russian assets for eventual forfeiture under new law (https://apnews.com/article/repo-act-banks-russia-ukraine-russian-assets-9ecda7e3e799cdbfb564844ae89a144b) Nonsense Darden Restaurants (NYSE: DRI) agreed to buy Tex-Mex chain Chuy's (https://www.axios.com/newsletters/axios-pro-rata-0bb0be2c-41d0-4181-bf39-ba3e827303da.html?chunk=1&utm_term=emshare#story1) Leadership within a costco warehouse (https://www.tiktok.com/t/ZTNaLxKJg/) If The Office took place at a car dealership (https://x.com/milkkarten/status/1813968113526067449?s=46&t=zgzybiDdIcGuQ_7WuoOX0A) Type in Morse code by repeatedly slamming your laptop shut (https://github.com/veggiedefender/open-and-shut) Sponsor SysAid – Next-Gen IT Service Management: (https://www.sysaid.com/lp/sysaid-copilot-l?utm_source=podcast&utm_medium=cpc&utm_campaign=software%20define) Experience the only platform with generative AI embedded in every aspect of IT management, enabling you to deliver exceptional service effortlessly and automagically. Listener Feedback (#asksdt) Foundation Models - IBM watsonx.ai (https://www.ibm.com/products/watsonx-ai/foundation-models) Conferences DevOpsDays Birmingham (https://devopsdays.org/events/2024-birmingham-al/welcome/), Aug 19-21, 2024 SpringOne (https://springone.io/?utm_source=cote&utm_campaign=devrel&utm_medium=newsletter&utm_content=newsletterUpcoming)/VMware Explore US (https://blogs.vmware.com/explore/2024/04/23/want-to-attend-vmware-explore-convince-your-manager-with-these/?utm_source=cote&utm_campaign=devrel&utm_medium=newsletter&utm_content=newsletterUpcoming), Aug 26-29, 2024 DevOpsDays Antwerp (https://devopsdays.org/events/2024-antwerp/welcome/), Sept 4–5, 2024, 15th anniversary SREday London 2024 (https://sreday.com/2024-london/), Sept 19–20, 2024 Coté speaking, 20% off with the code SRE20DAY (https://sreday.com/2024-london/#tickets) SDT News & Community Join our Slack community (https://softwaredefinedtalk.slack.com/join/shared_invite/zt-1hn55iv5d-UTfN7mVX1D9D5ExRt3ZJYQ#/shared-invite/email), post questions in #asksdt (https://softwaredefinedtalk.slack.com/archives/C07CSP19GAH) Email the show: questions@softwaredefinedtalk.com (mailto:questions@softwaredefinedtalk.com) Free stickers: Email your address to stickers@softwaredefinedtalk.com (mailto:stickers@softwaredefinedtalk.com) Follow us on social media: Twitter (https://twitter.com/softwaredeftalk), Threads (https://www.threads.net/@softwaredefinedtalk), Mastodon (https://hachyderm.io/@softwaredefinedtalk), LinkedIn (https://www.linkedin.com/company/software-defined-talk/), BlueSky (https://bsky.app/profile/softwaredefinedtalk.com) Watch us on: Twitch (https://www.twitch.tv/sdtpodcast), YouTube (https://www.youtube.com/channel/UCi3OJPV6h9tp-hbsGBLGsDQ/featured), Instagram (https://www.instagram.com/softwaredefinedtalk/), TikTok (https://www.tiktok.com/@softwaredefinedtalk) Book offer: Use code SDT for $20 off "Digital WTF" by Coté (https://leanpub.com/digitalwtf/c/sdt) Sponsor (https://www.softwaredefinedtalk.com/ads) the show (https://www.softwaredefinedtalk.com/ads) Recommendations Brandon: Austin FC (https://www.austinfc.com/competitions/mls-regular-season/2024/matches/atxvssea-07-13-2024/) Presumed Innocent (https://www.google.com/url?sa=t&source=web&rct=j&opi=89978449&url=https://tv.apple.com/us/show/presumed-innocent/umc.cmc.5hnqrhwtzt3esr7rb1wq2ppvn&ved=2ahUKEwiClKPk28CHAxWXLUQIHd59CCoQFnoECEcQAQ&usg=AOvVaw20AOPkQVtwWO77Jomxeua0) The Contrarian (https://www.penguinrandomhouse.com/books/609711/the-contrarian-by-max-chafkin/) Matt: Crying Out Cloud (https://www.wiz.io/crying-out-cloud) podcast Photo Credits Artwork (https://unsplash.com/photos/a-computer-screen-with-a-blue-screen-on-it-t_IkF_CNvSY)

Startup Hustle
Scaling an Enterprise Data Startup

Startup Hustle

Play Episode Listen Later Jul 18, 2024 39:37


In this episode of the Startup Hustle podcast, Matt Watson interviews Ozan Unlu, the founder and CEO of Edge Delta, an observability company. They discuss the challenges of managing large amounts of data in complex production environments and the need for developers to have access to production data without compromising security.  Ozan shares his journey from being a software developer at Microsoft to becoming a solution architect and eventually starting his own company. They also delve into the unique problem that Edge Delta solves, which is processing and analyzing data at the edge to determine its value and optimize storage and retention.  Takeaways- Observability is crucial for managing large amounts of data in complex production environments.- Developers need access to production data without compromising security.- Edge Delta solves the problem of processing and analyzing data at the edge to determine its value and optimize storage and retention.- The company provides flexible and scalable solutions for observability and data pipelines.- Edge Delta competes with log management tools like Splunk and DataDog, offering 80% of features that customers care about.- Edge Delta aims to be the best platform for customers in the long term, rather than focusing solely on revenue. Starting a deep tech and enterprise company requires a long-term vision and the ability to navigate both technical and sales challenges.- Raising early-stage funding can be lengthy, but having a strong team and unique insights can help attract investors.- The sales cycle for enterprise accounts is typically five to six months, involving multiple teams and evaluation periods.- Storing and analyzing large amounts of data in the observability space is a significant challenge, and implementing next-generation architectures can help bring down costs.- Entrepreneurs should seize the current market opportunities and the availability of talent and potential investors to pursue their dreams. Find Startup Hustle Everywhere:https://gigb.co/l/YEh5 This episode is sponsored by Full Scale:https://fullscale.io/ Visit the Edge Delta website:https://edgedelta.com/ Learn more about Ozan here:https://www.linkedin.com/in/ozanu/ Sign up for the Startup Hustle newsletter:https://newsletter.startuphustle.xyz/ Sound Bites"I'm excited because not very often do I get a podcast host who also has a tremendous amount of experience in the space.""I think my unique experience of being an ex-developer who then went solution architect and went salesperson, showing up once a week and really working with the customers on deep technical challenges, kind of in the front trenches, that...""Edge Delta is the concept of how do we start looking at data as it's being created, as close to the source of that data as possible, and how do we start to pre-process, pre-aggregate, and frankly have machine learning and some even basic AI running on the edge.""We have to have a very long-term vision on this, even on day one.""The team is probably the number one thing.""It feels like you're on the treadmill, just trying to get a few customers landed." Chapters00:00 The Challenges of Observability and Data Management06:17 Solving Unique Problems with Edge Delta12:55 Competing in the Landscape of Observability Vendors25:05 The Journey to Acquiring the First Paying Customer31:22 Innovation in the Observability SpaceSee omnystudio.com/listener for privacy information.

Venture Unlocked: The playbook for venture capital managers.
The great startup reckoning event of 2023 and 2024, but why startups should now start going back on offense featuring Tom Loverro of IVP

Venture Unlocked: The playbook for venture capital managers.

Play Episode Listen Later Jul 17, 2024 35:56


Follow me @samirkaji for my thoughts on the venture market, with a focus on the continued evolution of the VC landscape.Tom Loverro, General Partner at IVP is our guest as part of our Venture Unlocked Shorts series intended to go deep on a single topic.We revisit Tom's Twitter post from early 2023, which spoke to the market shift that was in motion and the difficulties start-ups would face in a capital-constrained market. Specifically, he spoke about 2024 as being a time of reckoning for many companies that were built with growth at all costs mentality. We went through that original post, and what's transpired since then, including why it's time for well-positioned startups to go on offense again. Tom brought a lot of interesting insights for founders and VCs alike, so we hope you enjoy the episode. About Tom Loverro:Tom Loverro is a General Partner at IVP in Menlo Park, California, where he focuses on investing in enterprise software and fintech companies. Since joining IVP in 2015, he has served as a Board Director or Observer for several companies, including Attentive, NerdWallet, Paper, Podium, Skydio, and TaxBit. He has also co-led investments in Amplitude, Datadog, GitHub, IEX, OnDeck, and Tanium.Prior to IVP, Tom was a Principal at RRE Ventures, focusing on early and mid-stage startups, and an Entrepreneur-in-Residence at Lightbank. He also served as Senior Director of Product Marketing at Drobo, Inc., and began his career as an Investment Banking Analyst at Goldman Sachs within the Technology, Media, and Telecommunications Group.Tom holds an MBA from the Kellogg School of Management at Northwestern University, with concentrations in Finance, Marketing, and Entrepreneurship & Innovation. He earned a BA in Political Science and History from Stanford University.In this episode, we discuss:(01:37) - Discussion on Tom's Twitter post from January 2023 and its context(02:09) - Tom's insights on the shift from a zero interest rate environment(02:59) - The concept of a mass extinction event for startups in 2023-2024(03:31) - Comparison with the Great Financial Crisis and its impact on startups(04:01) - The role of venture excess in 2021 and its aftermath(05:00) - Discussion on venture fund deployment and its impact on startups(06:49) - Dry powder theory and its implications on startup funding(07:49) - Insights on current market conditions and startup valuations(09:14) - Strategies startups adopted in response to market conditions(10:27) - The three archetypes of startups in the post-2021 era(13:18) - Observations on fundraising challenges and potential outcomes for startups(14:48) - Impact of LP capital dynamics on venture funding(16:34) - The evolving role of private equity in acquiring tech startups(18:09) - Comparison of venture fund impacts on early and late-stage investors(21:30) - Discussion on the IPO market and its high bar for startups(24:19) - The broader ecosystem of liquidity options for startups today(25:41) - Tom's recent post on shifting from defensive to offensive strategies(28:47) - Characteristics of startups that should consider going on offense(30:00) - Importance of survival, product-market fit, and unit economics for startups(31:50) - Potential exogenous events and their impact on market predictions(34:00) - Tom's advice to founders on acting with convictionI'd love to know what you took away from this conversation with Tom. Follow me @SamirKaji and give me your insights and questions with the hashtag #ventureunlocked. If you'd like to be considered as a guest or have someone you'd like to hear from (GP or LP), drop me a direct message on Twitter.Podcast Production support provided by Agent Bee This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit ventureunlocked.substack.com