Podcasts about ideogram

  • 61PODCASTS
  • 92EPISODES
  • 41mAVG DURATION
  • 1WEEKLY EPISODE
  • May 11, 2025LATEST

POPULARITY

20172018201920202021202220232024


Best podcasts about ideogram

Latest podcast episodes about ideogram

da Brand a Friend
#371 - Scimmie Anticocco

da Brand a Friend

Play Episode Listen Later May 11, 2025 21:18


#371 - Scimmie AnticoccoChiacchiere, riflessioni, pensieri. Nulla di speciale o straordinario. Quello che vedo, sento e mi passa per la testa.  Nuove mini-app create con l'IA1) WebPage TimeDetector - Chrome extension +2) Video Finder - web appdisponibili gratuitamente per tutti i miei sostenitori su Patreon e SubstackInstagram channel - momenti di vita non in posa - cosa vedono i miei occhi:https://instagram.com/giggi_canali _______________Info Utili• Sostieni questo podcast:Entra in contatto con me, ottieni feedback, ricevi consigli sul tuo progetto onlinehttps://Patreon.com/Robin_Good•  Musica di questa puntata:"Sleepyface" by Birocratic disponibile su Bandcamp•  Nella foto di copertina:Scimmia con cocco. Generata con Ideogram. • Ascolta e condividi questo podcast:https://www.spreaker.com/show/dabrandafriendArchivio completo organizzato per temi:https://start.me/p/kxENzk/da-brand-a-friend-archivio-podcast• Seguimi su Telegram:https://t.me/RobinGoodItalia• Newsletter in Inglese:https://robingood.substack.com - Fuoco su costruire fiducia per chi fa l'imprenditore onlinehttps://goodtools.substack.com - Tool alternativi a costo zerohttps://curationmonetized.substack.com - Esempi di come monetizzare organizzando informazioni.

Leveraging AI
187 | How Top Creators Use AI to Create Scroll-Stopping LinkedIn Content with MJ Jaindl

Leveraging AI

Play Episode Listen Later May 6, 2025 47:19 Transcription Available


Ryan's Method: Passive Income Podcast
Create Profitable Print on Demand Designs in 2 Minutes (or less)

Ryan's Method: Passive Income Podcast

Play Episode Listen Later Apr 15, 2025 8:55


One of the most popular design tools for print on demand just integrated Ideogram natively into their design app, allowing me to create 10 designs in 10 minutes!

Ryan's Method: Passive Income Podcast
Watch me Create Best-Selling Designs in 27 Seconds

Ryan's Method: Passive Income Podcast

Play Episode Listen Later Apr 2, 2025 14:17


Ideogram has an incredible ability to 'remix' existing designs and allow us to sell in the highest-demand print on demand niches with no graphic design experience or ability.

Let's Talk AI
#205 - Gemini 2.5, ChatGPT Image Gen, Thoughts of LLMs

Let's Talk AI

Play Episode Listen Later Apr 1, 2025 94:18 Transcription Available


Our 205th episode with a summary and discussion of last week's big AI news! Recorded on 03/28/2025 Hosted by Andrey Kurenkov and Jeremie Harris. Feel free to email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai Read out our text newsletter and comment on the podcast at https://lastweekin.ai/. Join our Discord here! https://discord.gg/nTyezGSKwP In this episode: OpenAI's new image generation capabilities represent significant advancements in AI tools, showcasing impressive benchmarks and multimodal functionalities. OpenAI is finalizing a historic $40 billion funding round led by SoftBank, and Sam Altman shifts focus to technical direction while COO Brad Lightcap takes on more operational responsibilities., Anthropic unveils groundbreaking interpretability research, introducing cross-layer tracers and showcasing deep insights into model reasoning through applications on Claude 3.5. New challenging benchmarks such as ARC AGI 2 and complex Sudoku variations aim to push the boundaries of reasoning and problem-solving capabilities in AI models. Timestamps + Links: (00:00:00) Intro / Banter (00:01:01) News Preview Tools & Apps (00:02:46) Gemini 2.5: Our most intelligent AI model (00:08:41) OpenAI rolls out image generation powered by GPT-4o to ChatGPT (00:16:14) Ideogram presents version 3.0 of its AI image generation system (00:19:20) New Reve Image Generator Beats AI Art Heavyweights MidJourney and Flux at a Penny Per Image (00:21:56) Alibaba Releases Qwen2.5 Omni, Adds Voice and Video Modes to Qwen Chat (00:23:58) The official version of Tencent's Hunyuan Deep Thinking Model T1 is here, with fast articulation, instant responses, and a decoding speed increase of 2 times Applications & Business (00:25:45) OpenAI Close to Finalizing $40 Billion SoftBank-Led Funding (00:29:26) OpenAI reshuffles leadership as Sam Altman pivots to technical focus (00:33:23) Nvidia shows off Rubin Ultra with 600,000-Watt Kyber racks and infrastructure, coming in 2027 (00:35:23) China's SiCarrier emerges as challenger to ASML, other chip tool titans (00:38:24) Pony.ai wins first permit for fully driverless taxi operation in the center of China's Silicon Valley Projects & Open Source (00:40:27) A new, challenging AGI test stumps most AI models (00:45:16) Challenging the Boundaries of Reasoning: An Olympiad-Level Math Benchmark for Large Language Models (00:48:13) Wan: Open and Advanced Large-Scale Video Generative Models (00:50:38) DeepSeek V3-0324 tops non-reasoning AI models in open-source first (00:54:46) OpenAI adopts rival Anthropic's standard for connecting AI models to data Research & Advancements (00:55:56) Anthropic can now track the bizarre inner workings of a large language model (01:06:00) Chain-of-Tools: Utilizing Massive Unseen Tools in the CoT Reasoning of Frozen Language Models (01:11:50) Inside-Out: Hidden Factual Knowledge in LLMs (01:15:14) Sakana AI super-powers AI reasoning using Japan's own Sudoku Puzzles Policy & Safety (01:18:38) Senator Wiener Introduces Legislation to Protect AI Whistleblowers & Boost Responsible AI Development (01:21:50) NVIDIA & Other Tech Giants Demand Trump Administration To Reconsider “AI Diffusion” Policy Which Is Set To Be Effective By May 15 (01:23:17) U.S. blacklists over 50 Chinese companies in bid to curb Beijing's AI, chip capabilities (01:26:44) Netflix's Reed Hastings Gives $50 Million to Bowdoin for A.I. Program (01:27:55) Judge allows 'New York Times' copyright case against OpenAI to go forward (01:29:48) Judge rules that AI can continue training on copyrighted lyrics, for now

Trench Tech
[Extrait] L'iA peut-elle sauver le monde ? - Lou Welgryn

Trench Tech

Play Episode Listen Later Mar 29, 2025 6:57


Et si l'IA n'était pas si magique que ça ? Parlons franchement : l'intelligence artificielle est-elle le problème ou la solution ?Ce podcast explore dégâts sociaux et environnementaux de l'IA et comment mobiliser citoyens et pros de la tech agir diffrémment. Ecoutez l'épisode complet Pirates de la Tech : Cap sur le bien commun avec Lou Welgryn :

VP Land
ChatGPT's Crazy Image Upgrade (Plus Reve & Ideogram 3.0), Hollywood's Global Shift, and Roblox Goes Generative

VP Land

Play Episode Listen Later Mar 28, 2025 34:58


We analyze the image generation capabilities of ChatGPT 4o, Reve, and Ideogram 3.0, examining their improved text handling and what it means for creators. Then, we dive into the Studio Ghibli AI controversy, Hollywood's production exodus overseas, and why Rob Lowe is shooting American game shows in Dublin. Plus, Roblox enters the generative 3D space with their new Cube tool, potentially changing how virtual worlds are built at scale.

TeknoSafari's Podcast
Okullarda AI Öğretmen Dönemi Başladı! Test Sonuçları Uçuşa Geçti

TeknoSafari's Podcast

Play Episode Listen Later Mar 28, 2025 22:54


1. Gemini, görüntü işlemede rekabeti kızıştırdı. Video analiz edebiliyor. 2.5 ile de başa oynuyor.2. GPT'ye yeni görüntü işleme motoru geldi pir geldi,3. Ideogram hemen Version 3'ü sürdü4. REVE çok çok iyi. 5. SORA, plus abonelere sınırsız6. GROK'a video üretme bekleniyor. 7. deepseek v3 geldi. R2 bekleniyor. Tst edenler AGI neredeyse diyor. Bu arada: deepseek bazı çalışanların yurtdışına serbestçe seyahat etmesini yasaklıyor.8. BAIDU, ERNIE 4.5 ve X1'i tanıttı.Çok modlu yeteneklere sahip derin düşünme akıl yürütme modeli olarak ERNIE X1, DeepSeek R1 ile aynı performansı yalnızca yarı fiyatına.9. QWEN'e video yüklenebiliyor.10. BYD Zhengzhou fabrikası San Francisco'dan DAHA BÜYÜK olacak. Tesla Gigafactory Nevada'dan 10 kat DAHA BÜYÜK.11. NotebookLM podcast özelliği Gemini'da. 12. NotebookLM'e mindmap geldi.13. Adobe'nin en iyi özelliklerini doğrudan iş akışınıza getiren yeni Microsoft 365 Copilot tanıtıldı.14. Adobe, üçüncü parti yazılımlara izin verecek. SORA ,Runway, Pika, Flux v.b. pek çok uygulama ekosisteme giriyor.15. Perplexity, Google ile dalga geçen reklam yayınladı. Cesur iş.16. Teksas okullarının ‘yapay zeka öğretmeni' kullanımı, öğrenci test puanlarını ülkedeki en iyi %2'ye fırlattı. Yöneticiler, öğrencilerin ‘daha iyi' ve ‘daha hızlı' öğrendiğini söylüyor.17. Steal Their Look adlı Glif, elbiselerinizi soyuyor. Glif.app, kullanıcıların AI tabanlı küçük uygulamalar ve sohbet botları oluşturmasına olanak tanıyan bir platform. Bu platformda oluşturulan "glifler", kullanıcı girdilerine (metin, resim veya buton tıklamaları) dayanarak metin, resim, video veya bunların kombinasyonlarını üreten AI destekli generatörlere verilen ad.18. Apple, NVIDIA GB300 NVL72 için 1 milyar dolarlık sipariş vererek yapay zeka veri merkezi oyununa adım atıyor. Şaka gibi.19. Nilüfer Belediyesi Yapay Zeka Bürosu'nu kuran ilk ilçe belediyesi oldu!

AI Unraveled: Latest AI News & Trends, Master GPT, Gemini, Generative AI, LLMs, Prompting, GPT Store

OpenAI faced copyright discussions over its Ghibli-style image generation while projecting substantial revenue growth, despite ongoing significant investment. Simultaneously, Ideogram launched a sophisticated image generation model, outperforming competitors. BMW and Alibaba partnered to integrate AI into vehicles, and Alibaba also released a versatile AI model for mobile devices and other applications, with plans for its adoption by major tech companies. Furthermore, Bill Gates predicted widespread replacement of doctors and teachers by AI, and North Korea revealed new AI-powered military drones, raising security considerations. The day also saw OpenAI enhance ChatGPT with image generation and adopt Anthropic's open-source protocol, alongside various other AI developments from companies like Microsoft, Amazon, and Midjourney, as well as regulatory actions.

AI For Humans
OpenAI's New 4o Image Gen Dominates The Internet, Google Gemini 2.5 & More AI News

AI For Humans

Play Episode Listen Later Mar 27, 2025 62:41


OpenAI's new 4o Image Gen is the best AI image model we've seen to date and it has absolutely taken over the Internet. Plus, Gemini 2.5 is no slouch and a ton of new robots! Plus, OpenAI's new OpenAI.fm let's you prompt AI voices in new ways, DeepSeek's new model is actually better (at times) then GPT 4.5, a new Cursor for 3D modeling and, we're so sorry for this, but a LOT of talk about AI Big Booty Bears.    **GO AND VISIT OUR SPONSOR Y'ALL** bubble.io/aiforhumans   Join the discord: https://discord.gg/muD2TYgC8f Join our Patreon: https://www.patreon.com/AIForHumansShow AI For Humans Newsletter: https://aiforhumans.beehiiv.com/ Follow us for more on X @AIForHumansShow Join our TikTok @aiforhumansshow To book us for speaking, please visit our website: https://www.aiforhumans.show/   // Show Links // OpenAI's GPT-4o Image Gen is Here https://openai.com/index/introducing-4o-image-generation/  Live demo (with Sam) https://www.youtube.com/live/2f3K43FHRKo?si=vL_0QC8ygRx4MgOF     OpenAI Causes The Great Giblification of the Internet https://x.com/heyBarsee/status/1904891940522647662 husbandt: https://x.com/squirtle_says/status/1904816587108213244 trump/vance: https://x.com/LukasMikelionis/status/1904873083246084364 movie scenes: https://x.com/MDurbar/status/1904872441899339963 brain meme: https://x.com/TechMemeKing/status/1904867629644267980 vibe ghibling: https://x.com/EMostaque/status/1904714479906283878   Sam Altman Says More Creative Freedom https://x.com/sama/status/1904598788687487422 Gavin's Knight + Rotisserie Chicken Photo Reddit Post https://www.reddit.com/r/ChatGPT/comments/1jk0p3v/tried_to_push_the_new_image_model_with_an/ Kevin's Aladdin Sane + Katamari WIlliams Images https://x.com/Attack/status/1904743185760608316 Big Butt Bear Video https://x.com/AIForHumansShow/status/1904687617758945674 Google Gemini 2.5  https://x.com/NoamShazeer/status/1904581813215125787 Largest Score Jump Ever on LMSYS https://x.com/AndrewCurran_/status/1904590242792996959 One Shot Coding Demos From Matt Berman https://x.com/MatthewBerman/status/1904714953095078004 Reve - Brand New Image Model Ranked #1 https://preview.reve.art/ Ideogram 3.0 https://x.com/ideogram_ai/status/1904927717281456188 OpenAI FM + new voice API https://x.com/OpenAIDevs/status/1902773579323674710  New DeepSeek Model is Actually Much Better https://www.reuters.com/technology/artificial-intelligence/chinas-deepseek-releases-ai-model-upgrade-intensifies-rivalry-with-openai-2025-03-25/ Figure 01 “Natural” Walking https://youtu.be/z6KiwXT_yAM?si=RRsmjvs0qpRU0cqX WPP Makes Robots Into Camera Operators https://x.com/TheHumanoidHub/status/1903173205155815431 H&M is making AI clones of 30 models https://www.inc.com/kit-eaton/clothing-giant-hm-will-use-models-ai-made-digital-twins-consent-included/91166352 Cursor for 3D Modeling  https://x.com/_martinsit/status/1904234440198615204 Seeing Eye Robot Dogs  https://x.com/iconphas/status/1904259348815352029 SynCity https://x.com/shtedritski/status/1903112129420443712 Gavin's Dial-up Diaries Video https://x.com/AIForHumansShow/status/1904244229783892207 Kevin's OpenAI Real Time Voice Test https://x.com/Attack/status/1904541254257643797      

Ryan's Method: Passive Income Podcast
IDEOGRAM 3.0 IS AMAZING FOR PRINT ON DEMAND!

Ryan's Method: Passive Income Podcast

Play Episode Listen Later Mar 27, 2025 11:12


I'm recreating 5 best-selling Etsy shirts using the new, powerful Ideogram 3.0 AI image generator!

Brave New Bookshelf
34 - Dana Sacco and Bootstrapping Your Publishing Career with AI Tools

Brave New Bookshelf

Play Episode Listen Later Mar 13, 2025 42:33


In this episode of Brave New Bookshelf, we sit down with Dana Sacco, a businesswoman-turned-author who has mastered the art of bootstrapping her publishing career using AI tools. Dana shares how she transitioned from being an avid reader to a multi-genre author, all while leveraging affordable and innovative AI solutions like ChatGPT, Claude, and Ideogram to streamline her workflow. Visit our website https://bravenewbookshelf.com to view the full episode notes, links and apps mentioned in the episode, and the full transcript.

Experts Unleashed with Joel Erway
Paid Ads Unleashed: Proven Strategies for Growth | EU 131 with Joe Stolte

Experts Unleashed with Joel Erway

Play Episode Listen Later Mar 12, 2025 29:19


In this episode of Experts Unleashed, I sit down with Joe Stolte from Daily AI to dive into his strategies for running successful paid ad campaigns. Joe shares how he spends around $15,000 a month on ads to drive free trials for his AI-powered email newsletter software, designed for thought leaders and small businesses. We talk about the power of retargeting, optimizing ad creatives, and using AI tools like Ideogram for image generation. Joe also highlights the value of partnering with social media creators to amplify results. Throughout our conversation, Joe emphasizes one key takeaway: having a strong offer and consistently testing are crucial for refining your marketing efforts.  

Ryan's Method: Passive Income Podcast
This Beginner-Friendly AI is Even Easier to Use Now

Ryan's Method: Passive Income Podcast

Play Episode Listen Later Mar 7, 2025 10:30


One of the most popular design tools for print on demand just integrated Ideogram natively into their design app, allowing me to create 10 designs in 10 minutes!

AI Unraveled: Latest AI News & Trends, Master GPT, Gemini, Generative AI, LLMs, Prompting, GPT Store

OpenAI released GPT-4.5, touting enhanced reasoning and emotional intelligence, while Tencent unveiled a rapid decision-making model aimed at real-time applications. Meta is developing a standalone AI app to compete with other chatbot platforms, and Amazon introduced its first quantum computing chip. Elsewhere, Ideogram is working on faster visual and textual processing, and Canada is scrutinising X's use of personal data for AI training. These developments, alongside anxieties about job displacement, signal a continued acceleration and broadening of AI's impact across various industries and societal domains.

The Post-Christian Podcast
A.I. for Church Leaders with Dr. David Thorne

The Post-Christian Podcast

Play Episode Listen Later Jan 14, 2025 21:42


In this episode of the Post Christian Podcast, host Dr. Eric Bryant interviews Dr. David Thorne from AI for Churches and leader of the AI Ethics Collective. From their website:"David earned his doctorate in Leadership Studies and masters degrees in management and leadership, counseling, and practical theology. Passionate about marketing strategy and helping churches utilize technology to achieve outreach goals and foster community engagement, David has a deep understanding of the hurdles church leaders face in a changing world."

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

Applications for the NYC AI Engineer Summit, focused on Agents at Work, are open!When we first started Latent Space, in the lightning round we'd always ask guests: “What's your favorite AI product?”. The majority would say Midjourney. The simple UI of prompt → very aesthetic image turned it into a $300M+ ARR bootstrapped business as it rode the first wave of AI image generation.In open source land, StableDiffusion was congregating around AUTOMATIC1111 as the de-facto web UI. Unlike Midjourney, which offered some flags but was mostly prompt-driven, A1111 let users play with a lot more parameters, supported additional modalities like img2img, and allowed users to load in custom models. If you're interested in some of the SD history, you can look at our episodes with Lexica, Replicate, and Playground.One of the people involved with that community was comfyanonymous, who was also part of the Stability team in 2023, decided to build an alternative called ComfyUI, now one of the fastest growing open source projects in generative images, and is now the preferred partner for folks like Black Forest Labs's Flux Tools on Day 1. The idea behind it was simple: “Everyone is trying to make easy to use interfaces. Let me try to make a powerful interface that's not easy to use.”Unlike its predecessors, ComfyUI does not have an input text box. Everything is based around the idea of a node: there's a text input node, a CLIP node, a checkpoint loader node, a KSampler node, a VAE node, etc. While daunting for simple image generation, the tool is amazing for more complex workflows since you can break down every step of the process, and then chain many of them together rather than manually switching between tools. You can also re-start execution halfway instead of from the beginning, which can save a lot of time when using larger models.To give you an idea of some of the new use cases that this type of UI enables:* Sketch something → Generate an image with SD from sketch → feed it into SD Video to animate* Generate an image of an object → Turn into a 3D asset → Feed into interactive experiences* Input audio → Generate audio-reactive videosTheir Examples page also includes some of the more common use cases like AnimateDiff, etc. They recently launched the Comfy Registry, an online library of different nodes that users can pull from rather than having to build everything from scratch. The project has >60,000 Github stars, and as the community grows, some of the projects that people build have gotten quite complex:The most interesting thing about Comfy is that it's not a UI, it's a runtime. You can build full applications on top of image models simply by using Comfy. You can expose Comfy workflows as an endpoint and chain them together just like you chain a single node. We're seeing the rise of AI Engineering applied to art.Major Tom's ComfyUI Resources from the Latent Space DiscordMajor shoutouts to Major Tom on the LS Discord who is a image generation expert, who offered these pointers:* “best thing about comfy is the fact it supports almost immediately every new thing that comes out - unlike A1111 or forge, which still don't support flux cnet for instance. It will be perfect tool when conflicting nodes will be resolved”* AP Workflows from Alessandro Perili are a nice example of an all-in-one train-evaluate-generate system built atop Comfy* ComfyUI YouTubers to learn from:* @sebastiankamph* @NerdyRodent* @OlivioSarikas* @sedetweiler* @pixaroma* ComfyUI Nodes to check out:* https://github.com/kijai/ComfyUI-IC-Light* https://github.com/MrForExample/ComfyUI-3D-Pack* https://github.com/PowerHouseMan/ComfyUI-AdvancedLivePortrait* https://github.com/pydn/ComfyUI-to-Python-Extension* https://github.com/THtianhao/ComfyUI-Portrait-Maker* https://github.com/ssitu/ComfyUI_NestedNodeBuilder* https://github.com/longgui0318/comfyui-magic-clothing* https://github.com/atmaranto/ComfyUI-SaveAsScript* https://github.com/ZHO-ZHO-ZHO/ComfyUI-InstantID* https://github.com/AIFSH/ComfyUI-FishSpeech* https://github.com/coolzilj/ComfyUI-Photopea* https://github.com/lks-ai/anynode* Sarav: https://www.youtube.com/@mickmumpitz/videos ( applied stuff )* Sarav: https://www.youtube.com/@latentvision (technical, but infrequent)* look for comfyui node for https://github.com/magic-quill/MagicQuill* “Comfy for Video” resources* Kijai (https://github.com/kijai) pushing out support for Mochi, CogVideoX, AnimateDif, LivePortrait etc* Comfyui node support like LTX https://github.com/Lightricks/ComfyUI-LTXVideo , and HunyuanVideo* FloraFauna AI* Communities: https://www.reddit.com/r/StableDiffusion/, https://www.reddit.com/r/comfyui/Full YouTube EpisodeAs usual, you can find the full video episode on our YouTube (and don't forget to like and subscribe!)Timestamps* 00:00:04 Introduction of hosts and anonymous guest* 00:00:35 Origins of Comfy UI and early Stable Diffusion landscape* 00:02:58 Comfy's background and development of high-res fix* 00:05:37 Area conditioning and compositing in image generation* 00:07:20 Discussion on different AI image models (SD, Flux, etc.)* 00:11:10 Closed source model APIs and community discussions on SD versions* 00:14:41 LoRAs and textual inversion in image generation* 00:18:43 Evaluation methods in the Comfy community* 00:20:05 CLIP models and text encoders in image generation* 00:23:05 Prompt weighting and negative prompting* 00:26:22 Comfy UI's unique features and design choices* 00:31:00 Memory management in Comfy UI* 00:33:50 GPU market share and compatibility issues* 00:35:40 Node design and parameter settings in Comfy UI* 00:38:44 Custom nodes and community contributions* 00:41:40 Video generation models and capabilities* 00:44:47 Comfy UI's development timeline and rise to popularity* 00:48:13 Current state of Comfy UI team and future plans* 00:50:11 Discussion on other Comfy startups and potential text generation supportTranscriptAlessio [00:00:04]: Hey everyone, welcome to the Latent Space podcast. This is Alessio, partner and CTO at Decibel Partners, and I'm joined by my co-host Swyx, founder of Small AI.swyx [00:00:12]: Hey everyone, we are in the Chroma Studio again, but with our first ever anonymous guest, Comfy Anonymous, welcome.Comfy [00:00:19]: Hello.swyx [00:00:21]: I feel like that's your full name, you just go by Comfy, right?Comfy [00:00:24]: Yeah, well, a lot of people just call me Comfy, even when they know my real name. Hey, Comfy.Alessio [00:00:32]: Swyx is the same. You know, not a lot of people call you Shawn.swyx [00:00:35]: Yeah, you have a professional name, right, that people know you by, and then you have a legal name. Yeah, it's fine. How do I phrase this? I think people who are in the know, know that Comfy is like the tool for image generation and now other multimodality stuff. I would say that when I first got started with Stable Diffusion, the star of the show was Automatic 111, right? And I actually looked back at my notes from 2022-ish, like Comfy was already getting started back then, but it was kind of like the up and comer, and your main feature was the flowchart. Can you just kind of rewind to that moment, that year and like, you know, how you looked at the landscape there and decided to start Comfy?Comfy [00:01:10]: Yeah, I discovered Stable Diffusion in 2022, in October 2022. And, well, I kind of started playing around with it. Yes, I, and back then I was using Automatic, which was what everyone was using back then. And so I started with that because I had, it was when I started, I had no idea like how Diffusion works. I didn't know how Diffusion models work, how any of this works, so.swyx [00:01:36]: Oh, yeah. What was your prior background as an engineer?Comfy [00:01:39]: Just a software engineer. Yeah. Boring software engineer.swyx [00:01:44]: But like any, any image stuff, any orchestration, distributed systems, GPUs?Comfy [00:01:49]: No, I was doing basically nothing interesting. Crud, web development? Yeah, a lot of web development, just, yeah, some basic, maybe some basic like automation stuff. Okay. Just. Yeah, no, like, no big companies or anything.swyx [00:02:08]: Yeah, but like already some interest in automations, probably a lot of Python.Comfy [00:02:12]: Yeah, yeah, of course, Python. But I wasn't actually used to like the Node graph interface before I started Comfy UI. It was just, I just thought it was like, oh, like, what's the best way to represent the Diffusion process in the user interface? And then like, oh, well. Well, like, naturally, oh, this is the best way I've found. And this was like with the Node interface. So how I got started was, yeah, so basic October 2022, just like I hadn't written a line of PyTorch before that. So it's completely new. What happened was I kind of got addicted to generating images.Alessio [00:02:58]: As we all did. Yeah.Comfy [00:03:00]: And then I started. I started experimenting with like the high-res fixed in auto, which was for those that don't know, the high-res fix is just since the Diffusion models back then could only generate that low-resolution. So what you would do, you would generate low-resolution image, then upscale, then refine it again. And that was kind of the hack to generate high-resolution images. I really liked generating. Like higher resolution images. So I was experimenting with that. And so I modified the code a bit. Okay. What happens if I, if I use different samplers on the second pass, I was edited the code of auto. So what happens if I use a different sampler? What happens if I use a different, like a different settings, different number of steps? And because back then the. The high-res fix was very basic, just, so. Yeah.swyx [00:04:05]: Now there's a whole library of just, uh, the upsamplers.Comfy [00:04:08]: I think, I think they added a bunch of, uh, of options to the high-res fix since, uh, since, since then. But before that was just so basic. So I wanted to go further. I wanted to try it. What happens if I use a different model for the second, the second pass? And then, well, then the auto code base was, wasn't good enough for. Like, it would have been, uh, harder to implement that in the auto interface than to create my own interface. So that's when I decided to create my own. And you were doing that mostly on your own when you started, or did you already have kind of like a subgroup of people? No, I was, uh, on my own because, because it was just me experimenting with stuff. So yeah, that was it. Then, so I started writing the code January one. 2023, and then I released the first version on GitHub, January 16th, 2023. That's how things got started.Alessio [00:05:11]: And what's, what's the name? Comfy UI right away or? Yeah.Comfy [00:05:14]: Comfy UI. The reason the name, my name is Comfy is people thought my pictures were comfy, so I just, uh, just named it, uh, uh, it's my Comfy UI. So yeah, that's, uh,swyx [00:05:27]: Is there a particular segment of the community that you targeted as users? Like more intensive workflow artists, you know, compared to the automatic crowd or, you know,Comfy [00:05:37]: This was my way of like experimenting with, uh, with new things, like the high risk fixed thing I mentioned, which was like in Comfy, the first thing you could easily do was just chain different models together. And then one of the first things, I think the first times it got a bit of popularity was when I started experimenting with the different, like applying. Prompts to different areas of the image. Yeah. I called it area conditioning, posted it on Reddit and it got a bunch of upvotes. So I think that's when, like, when people first learned of Comfy UI.swyx [00:06:17]: Is that mostly like fixing hands?Comfy [00:06:19]: Uh, no, no, no. That was just, uh, like, let's say, well, it was very, well, it still is kind of difficult to like, let's say you want a mountain, you have an image and then, okay. I'm like, okay. I want the mountain here and I want the, like a, a Fox here.swyx [00:06:37]: Yeah. So compositing the image. Yeah.Comfy [00:06:40]: My way was very easy. It was just like, oh, when you run the diffusion process, you kind of generate, okay. You do pass one pass through the diffusion, every step you do one pass. Okay. This place of the image with this brand, this space, place of the image with the other prop. And then. The entire image with another prop and then just average everything together, every step, and that was, uh, area composition, which I call it. And then, then a month later, there was a paper that came out called multi diffusion, which was the same thing, but yeah, that's, uh,Alessio [00:07:20]: could you do area composition with different models or because you're averaging out, you kind of need the same model.Comfy [00:07:26]: Could do it with, but yeah, I hadn't implemented it. For different models, but, uh, you, you can do it with, uh, with different models if you want, as long as the models share the same latent space, like we, we're supposed to ring a bell every time someone says, yeah, like, for example, you couldn't use like Excel and SD 1.5, because those have a different latent space, but like, uh, yeah, like SD 1.5 models, different ones. You could, you could do that.swyx [00:07:59]: There's some models that try to work in pixel space, right?Comfy [00:08:03]: Yeah. They're very slow. Of course. That's the problem. That that's the, the reason why stable diffusion actually became like popular, like, cause was because of the latent space.swyx [00:08:14]: Small and yeah. Because it used to be latent diffusion models and then they trained it up.Comfy [00:08:19]: Yeah. Cause a pixel pixel diffusion models are just too slow. So. Yeah.swyx [00:08:25]: Have you ever tried to talk to like, like stability, the latent diffusion guys, like, you know, Robin Rombach, that, that crew. Yeah.Comfy [00:08:32]: Well, I used to work at stability.swyx [00:08:34]: Oh, I actually didn't know. Yeah.Comfy [00:08:35]: I used to work at stability. I got, uh, I got hired, uh, in June, 2023.swyx [00:08:42]: Ah, that's the part of the story I didn't know about. Okay. Yeah.Comfy [00:08:46]: So the, the reason I was hired is because they were doing, uh, SDXL at the time and they were basically SDXL. I don't know if you remember it was a base model and then a refiner model. Basically they wanted to experiment, like chaining them together. And then, uh, they saw, oh, right. Oh, this, we can use this to do that. Well, let's hire that guy.swyx [00:09:10]: But they didn't, they didn't pursue it for like SD3. What do you mean? Like the SDXL approach. Yeah.Comfy [00:09:16]: The reason for that approach was because basically they had two models and then they wanted to publish both of them. So they, they trained one on. Lower time steps, which was the refiner model. And then they, the first one was trained normally. And then they went during their test, they realized, oh, like if we string these models together are like quality increases. So let's publish that. It worked. Yeah. But like right now, I don't think many people actually use the refiner anymore, even though it is actually a full diffusion model. Like you can use it on its own. And it's going to generate images. I don't think anyone, people have mostly forgotten about it. But, uh.Alessio [00:10:05]: Can we talk about models a little bit? So stable diffusion, obviously is the most known. I know flux has gotten a lot of traction. Are there any underrated models that people should use more or what's the state of the union?Comfy [00:10:17]: Well, the, the latest, uh, state of the art, at least, yeah, for images there's, uh, yeah, there's flux. There's also SD3.5. SD3.5 is two models. There's a, there's a small one, 2.5B and there's the bigger one, 8B. So it's, it's smaller than flux. So, and it's more, uh, creative in a way, but flux, yeah, flux is the best. People should give SD3.5 a try cause it's, uh, it's different. I won't say it's better. Well, it's better for some like specific use cases. Right. If you want some to make something more like creative, maybe SD3.5. If you want to make something more consistent and flux is probably better.swyx [00:11:06]: Do you ever consider supporting the closed source model APIs?Comfy [00:11:10]: Uh, well, they, we do support them as custom nodes. We actually have some, uh, official custom nodes from, uh, different. Ideogram.swyx [00:11:20]: Yeah. I guess DALI would have one. Yeah.Comfy [00:11:23]: That's, uh, it's just not, I'm not the person that handles that. Sure.swyx [00:11:28]: Sure. Quick question on, on SD. There's a lot of community discussion about the transition from SD1.5 to SD2 and then SD2 to SD3. People still like, you know, very loyal to the previous generations of SDs?Comfy [00:11:41]: Uh, yeah. SD1.5 then still has a lot of, a lot of users.swyx [00:11:46]: The last based model.Comfy [00:11:49]: Yeah. Then SD2 was mostly ignored. It wasn't, uh, it wasn't a big enough improvement over the previous one. Okay.swyx [00:11:58]: So SD1.5, SD3, flux and whatever else. SDXL. SDXL.Comfy [00:12:03]: That's the main one. Stable cascade. Stable cascade. That was a good model. But, uh, that's, uh, the problem with that one is, uh, it got, uh, like SD3 was announced one week after. Yeah.swyx [00:12:16]: It was like a weird release. Uh, what was it like inside of stability actually? I mean, statute of limitations. Yeah. The statute of limitations expired. You know, management has moved. So it's easier to talk about now. Yeah.Comfy [00:12:27]: And inside stability, actually that model was ready, uh, like three months before, but it got, uh, stuck in, uh, red teaming. So basically the product, if that model had released or was supposed to be released by the authors, then it would probably have gotten very popular since it's a, it's a step up from SDXL. But it got all of its momentum stolen. It got stolen by the SD3 announcement. So people kind of didn't develop anything on top of it, even though it's, uh, yeah. It was a good model, at least, uh, completely mostly ignored for some reason. Likeswyx [00:13:07]: I think the naming as well matters. It seemed like a branch off of the main, main tree of development. Yeah.Comfy [00:13:15]: Well, it was different researchers that did it. Yeah. Yeah. Very like, uh, good model. Like it's the Worcestershire authors. I don't know if I'm pronouncing it correctly. Yeah. Yeah. Yeah.swyx [00:13:28]: I actually met them in Vienna. Yeah.Comfy [00:13:30]: They worked at stability for a bit and they left right after the Cascade release.swyx [00:13:35]: This is Dustin, right? No. Uh, Dustin's SD3. Yeah.Comfy [00:13:38]: Dustin is a SD3 SDXL. That's, uh, Pablo and Dome. I think I'm pronouncing his name correctly. Yeah. Yeah. Yeah. Yeah. That's very good.swyx [00:13:51]: It seems like the community is very, they move very quickly. Yeah. Like when there's a new model out, they just drop whatever the current one is. And they just all move wholesale over. Like they don't really stay to explore the full capabilities. Like if, if the stable cascade was that good, they would have AB tested a bit more. Instead they're like, okay, SD3 is out. Let's go. You know?Comfy [00:14:11]: Well, I find the opposite actually. The community doesn't like, they only jump on a new model when there's a significant improvement. Like if there's a, only like a incremental improvement, which is what, uh, most of these models are going to have, especially if you, cause, uh, stay the same parameter count. Yeah. Like you're not going to get a massive improvement, uh, into like, unless there's something big that, that changes. So, uh. Yeah.swyx [00:14:41]: And how are they evaluating these improvements? Like, um, because there's, it's a whole chain of, you know, comfy workflows. Yeah. How does, how does one part of the chain actually affect the whole process?Comfy [00:14:52]: Are you talking on the model side specific?swyx [00:14:54]: Model specific, right? But like once you have your whole workflow based on a model, it's very hard to move.Comfy [00:15:01]: Uh, not, well, not really. Well, it depends on your, uh, depends on their specific kind of the workflow. Yeah.swyx [00:15:09]: So I do a lot of like text and image. Yeah.Comfy [00:15:12]: When you do change, like most workflows are kind of going to be complete. Yeah. It's just like, you might have to completely change your prompt completely change. Okay.swyx [00:15:24]: Well, I mean, then maybe the question is really about evals. Like what does the comfy community do for evals? Just, you know,Comfy [00:15:31]: Well, that they don't really do that. It's more like, oh, I think this image is nice. So that's, uh,swyx [00:15:38]: They just subscribe to Fofr AI and just see like, you know, what Fofr is doing. Yeah.Comfy [00:15:43]: Well, they just, they just generate like it. Like, I don't see anyone really doing it. Like, uh, at least on the comfy side, comfy users, they, it's more like, oh, generate images and see, oh, this one's nice. It's like, yeah, it's not, uh, like the, the more, uh, like, uh, scientific, uh, like, uh, like checking that's more on specifically on like model side. If, uh, yeah, but there is a lot of, uh, vibes also, cause it is a like, uh, artistic, uh, you can create a very good model that doesn't generate nice images. Cause most images on the internet are ugly. So if you, if that's like, if you just, oh, I have the best model at 10th giant, it's super smart. I created on all the, like I've trained on just all the images on the internet. The images are not going to look good. So yeah.Alessio [00:16:42]: Yeah.Comfy [00:16:43]: They're going to be very consistent. But yeah. People like, it's not going to be like the, the look that people are going to be expecting from, uh, from a model. So. Yeah.swyx [00:16:54]: Can we talk about LoRa's? Cause we thought we talked about models then like the next step is probably LoRa's. Before, I actually, I'm kind of curious how LoRa's entered the tool set of the image community because the LoRa paper was 2021. And then like, there was like other methods like textual inversion that was popular at the early SD stage. Yeah.Comfy [00:17:13]: I can't even explain the difference between that. Yeah. Textual inversions. That's basically what you're doing is you're, you're training a, cause well, yeah. Stable diffusion. You have the diffusion model, you have text encoder. So basically what you're doing is training a vector that you're going to pass to the text encoder. It's basically you're training a new word. Yeah.swyx [00:17:37]: It's a little bit like representation engineering now. Yeah.Comfy [00:17:40]: Yeah. Basically. Yeah. You're just, so yeah, if you know how like the text encoder works, basically you have, you take your, your words of your product, you convert those into tokens with the tokenizer and those are converted into vectors. Basically. Yeah. Each token represents a different vector. So each word presents a vector. And those, depending on your words, that's the list of vectors that get passed to the text encoder, which is just. Yeah. Yeah. I'm just a stack of, of attention. Like basically it's a very close to LLM architecture. Yeah. Yeah. So basically what you're doing is just training a new vector. We're saying, well, I have all these images and I want to know which word does that represent? And it's going to get like, you train this vector and then, and then when you use this vector, it hopefully generates. Like something similar to your images. Yeah.swyx [00:18:43]: I would say it's like surprisingly sample efficient in picking up the concept that you're trying to train it on. Yeah.Comfy [00:18:48]: Well, people have kind of stopped doing that even though back as like when I was at Stability, we, we actually did train internally some like textual versions on like T5 XXL actually worked pretty well. But for some reason, yeah, people don't use them. And also they might also work like, like, yeah, this is something and probably have to test, but maybe if you train a textual version, like on T5 XXL, it might also work with all the other models that use T5 XXL because same thing with like, like the textual inversions that, that were trained for SD 1.5, they also kind of work on SDXL because SDXL has the, has two text encoders. And one of them is the same as the, as the SD 1.5 CLIP-L. So those, they actually would, they don't work as strongly because they're only applied to one of the text encoders. But, and the same thing for SD3. SD3 has three text encoders. So it works. It's still, you can still use your textual version SD 1.5 on SD3, but it's just a lot weaker because now there's three text encoders. So it gets even more diluted. Yeah.swyx [00:20:05]: Do people experiment a lot on, just on the CLIP side, there's like Siglip, there's Blip, like do people experiment a lot on those?Comfy [00:20:12]: You can't really replace. Yeah.swyx [00:20:14]: Because they're trained together, right? Yeah.Comfy [00:20:15]: They're trained together. So you can't like, well, what I've seen people experimenting with is a long CLIP. So basically someone fine tuned the CLIP model to accept longer prompts.swyx [00:20:27]: Oh, it's kind of like long context fine tuning. Yeah.Comfy [00:20:31]: So, so like it's, it's actually supported in Core Comfy.swyx [00:20:35]: How long is long?Comfy [00:20:36]: Regular CLIP is 77 tokens. Yeah. Long CLIP is 256. Okay. So, but the hack that like you've, if you use stable diffusion 1.5, you've probably noticed, oh, it still works if I, if I use long prompts, prompts longer than 77 words. Well, that's because the hack is to just, well, you split, you split it up in chugs of 77, your whole big prompt. Let's say you, you give it like the massive text, like the Bible or something, and it would split it up in chugs of 77 and then just pass each one through the CLIP and then just cut anything together at the end. It's not ideal, but it actually works.swyx [00:21:26]: Like the positioning of the words really, really matters then, right? Like this is why order matters in prompts. Yeah.Comfy [00:21:33]: Yeah. Like it, it works, but it's, it's not ideal, but it's what people expect. Like if, if someone gives a huge prompt, they expect at least some of the concepts at the end to be like present in the image. But usually when they give long prompts, they, they don't, they like, they don't expect like detail, I think. So that's why it works very well.swyx [00:21:58]: And while we're on this topic, prompts waiting, negative comments. Negative prompting all, all sort of similar part of this layer of the stack. Yeah.Comfy [00:22:05]: The, the hack for that, which works on CLIP, like it, basically it's just for SD 1.5, well, for SD 1.5, the prompt waiting works well because CLIP L is a, is not a very deep model. So you have a very high correlation between, you have the input token, the index of the input token vector. And the output token, they're very, the concepts are very close, closely linked. So that means if you interpolate the vector from what, well, the, the way Comfy UI does it is it has, okay, you have the vector, you have an empty prompt. So you have a, a chunk, like a CLIP output for the empty prompt, and then you have the one for your prompt. And then it interpolates from that, depending on your prompt. Yeah.Comfy [00:23:07]: So that's how it, how it does prompt waiting. But this stops working the deeper your text encoder is. So on T5X itself, it doesn't work at all. So. Wow.swyx [00:23:20]: Is that a problem for people? I mean, cause I'm used to just move, moving up numbers. Probably not. Yeah.Comfy [00:23:25]: Well.swyx [00:23:26]: So you just use words to describe, right? Cause it's a bigger language model. Yeah.Comfy [00:23:30]: Yeah. So. Yeah. So honestly it might be good, but I haven't seen many complaints on Flux that it's not working. So, cause I guess people can sort of get around it with, with language. So. Yeah.swyx [00:23:46]: Yeah. And then coming back to LoRa's, now the, the popular way to, to customize models is LoRa's. And I saw you also support Locon and LoHa, which I've never heard of before.Comfy [00:23:56]: There's a bunch of, cause what, what the LoRa is essentially is. Instead of like, okay, you have your, your model and then you want to fine tune it. So instead of like, what you could do is you could fine tune the entire thing, but that's a bit heavy. So to speed things up and make things less heavy, what you can do is just fine tune some smaller weights, like basically two, two matrices that when you multiply like two low rank matrices and when you multiply them together, gives a, represents a difference between trained weights and your base weights. So by training those two smaller matrices, that's a lot less heavy. Yeah.Alessio [00:24:45]: And they're portable. So you're going to share them. Yeah. It's like easier. And also smaller.Comfy [00:24:49]: Yeah. That's the, how LoRa's work. So basically, so when, when inferencing you, you get an inference with them pretty efficiently, like how ComputeWrite does it. It just, when you use a LoRa, it just applies it straight on the weights so that there's only a small delay at the base, like before the sampling to when it applies the weights and then it just same speed as, as before. So for, for inference, it's, it's not that bad, but, and then you have, so basically all the LoRa types like LoHa, LoCon, everything, that's just different ways of representing that like. Basically, you can call it kind of like compression, even though it's not really compression, it's just different ways of represented, like just, okay, I want to train a different on the difference on the weights. What's the best way to represent that difference? There's the basic LoRa, which is just, oh, let's multiply these two matrices together. And then there's all the other ones, which are all different algorithms. So. Yeah.Alessio [00:25:57]: So let's talk about LoRa. Let's talk about what comfy UI actually is. I think most people have heard of it. Some people might've seen screenshots. I think fewer people have built very complex workflows. So when you started, automatic was like the super simple way. What were some of the choices that you made? So the node workflow, is there anything else that stands out as like, this was like a unique take on how to do image generation workflows?Comfy [00:26:22]: Well, I feel like, yeah, back then everyone was trying to make like easy to use interface. Yeah. So I'm like, well, everyone's trying to make an easy to use interface.swyx [00:26:32]: Let's make a hard to use interface.Comfy [00:26:37]: Like, so like, I like, I don't need to do that, everyone else doing it. So let me try something like, let me try to make a powerful interface that's not easy to use. So.swyx [00:26:52]: So like, yeah, there's a sort of node execution engine. Yeah. Yeah. And it actually lists, it has this really good list of features of things you prioritize, right? Like let me see, like sort of re-executing from, from any parts of the workflow that was changed, asynchronous queue system, smart memory management, like all this seems like a lot of engineering that. Yeah.Comfy [00:27:12]: There's a lot of engineering in the back end to make things, cause I was always focused on making things work locally very well. Cause that's cause I was using it locally. So everything. So there's a lot of, a lot of thought and working by getting everything to run as well as possible. So yeah. ConfUI is actually more of a back end, at least, well, not all the front ends getting a lot more development, but, but before, before it was, I was pretty much only focused on the backend. Yeah.swyx [00:27:50]: So v0.1 was only August this year. Yeah.Comfy [00:27:54]: With the new front end. Before there was no versioning. So yeah. Yeah. Yeah.swyx [00:27:57]: And so what was the big rewrite for the 0.1 and then the 1.0?Comfy [00:28:02]: Well, that's more on the front end side. That's cause before that it was just like the UI, what, cause when I first wrote it, I just, I said, okay, how can I make, like, I can do web development, but I don't like doing it. Like what's the easiest way I can slap a node interface on this. And then I found this library. Yeah. Like JavaScript library.swyx [00:28:26]: Live graph?Comfy [00:28:27]: Live graph.swyx [00:28:28]: Usually people will go for like react flow for like a flow builder. Yeah.Comfy [00:28:31]: But that seems like too complicated. So I didn't really want to spend time like developing the front end. So I'm like, well, oh, light graph. This has the whole node interface. So, okay. Let me just plug that into, to my backend.swyx [00:28:49]: I feel like if Streamlit or Gradio offered something that you would have used Streamlit or Gradio cause it's Python. Yeah.Comfy [00:28:54]: Yeah. Yeah. Yeah.Comfy [00:29:00]: Yeah.Comfy [00:29:14]: Yeah. logic and your backend logic and just sticks them together.swyx [00:29:20]: It's supposed to be easy for you guys. If you're a Python main, you know, I'm a JS main, right? Okay. If you're a Python main, it's supposed to be easy.Comfy [00:29:26]: Yeah, it's easy, but it makes your whole software a huge mess.swyx [00:29:30]: I see, I see. So you're mixing concerns instead of separating concerns?Comfy [00:29:34]: Well, it's because... Like frontend and backend. Frontend and backend should be well separated with a defined API. Like that's how you're supposed to do it. Smart people disagree. It just sticks everything together. It makes it easy to like a huge mess. And also it's, there's a lot of issues with Gradio. Like it's very good if all you want to do is just get like slap a quick interface on your, like to show off your ML project. Like that's what it's made for. Yeah. Like there's no problem using it. Like, oh, I have my, I have my code. I just wanted a quick interface on it. That's perfect. Like use Gradio. But if you want to make something that's like a real, like real software that will last a long time and will be easy to maintain, then I would avoid it. Yeah.swyx [00:30:32]: So your criticism is Streamlit and Gradio are the same. I mean, those are the same criticisms.Comfy [00:30:37]: Yeah, Streamlit I haven't used as much. Yeah, I just looked a bit.swyx [00:30:43]: Similar philosophy.Comfy [00:30:44]: Yeah, it's similar. It's just, it just seems to me like, okay, for quick, like AI demos, it's perfect.swyx [00:30:51]: Yeah. Going back to like the core tech, like asynchronous queues, slow re-execution, smart memory management, you know, anything that you were very proud of or was very hard to figure out?Comfy [00:31:00]: Yeah. The thing that's the biggest pain in the ass is probably the memory management. Yeah.swyx [00:31:05]: Were you just paging models in and out or? Yeah.Comfy [00:31:08]: Before it was just, okay, load the model, completely unload it. Then, okay, that, that works well when you, your model are small, but if your models are big and it takes sort of like, let's say someone has a, like a, a 4090, and the model size is 10 gigabytes, that can take a few seconds to like load and load, load and load, so you want to try to keep things like in memory, in the GPU memory as much as possible. What Comfy UI does right now is it. It tries to like estimate, okay, like, okay, you're going to sample this model, it's going to take probably this amount of memory, let's remove the models, like this amount of memory that's been loaded on the GPU and then just execute it. But so there's a fine line between just because try to remove the least amount of models that are already loaded. Because as fans, like Windows drivers, and one other problem is the NVIDIA driver on Windows by default, because there's a way to, there's an option to disable that feature, but by default it, like, if you start loading, you can overflow your GPU memory and then it's, the driver's going to automatically start paging to RAM. But the problem with that is it's, it makes everything extremely slow. So when you see people complaining, oh, this model, it works, but oh, s**t, it starts slowing down a lot, that's probably what's happening. So it's basically you have to just try to get, use as much memory as possible, but not too much, or else things start slowing down, or people get out of memory, and then just find, try to find that line where, oh, like the driver on Windows starts paging and stuff. Yeah. And the problem with PyTorch is it's, it's high levels, don't have that much fine-grained control over, like, specific memory stuff, so kind of have to leave, like, the memory freeing to, to Python and PyTorch, which is, can be annoying sometimes.swyx [00:33:32]: So, you know, I think one thing is, as a maintainer of this project, like, you're designing for a very wide surface area of compute, like, you even support CPUs.Comfy [00:33:42]: Yeah, well, that's... That's just, for PyTorch, PyTorch supports CPUs, so, yeah, it's just, that's not, that's not hard to support.swyx [00:33:50]: First of all, is there a market share estimate, like, is it, like, 70% NVIDIA, like, 30% AMD, and then, like, miscellaneous on Apple, Silicon, or whatever?Comfy [00:33:59]: For Comfy? Yeah. Yeah, and, yeah, I don't know the market share.swyx [00:34:03]: Can you guess?Comfy [00:34:04]: I think it's mostly NVIDIA. Right. Because, because AMD, the problem, like, AMD works horribly on Windows. Like, on Linux, it works fine. It's, it's lower than the price equivalent NVIDIA GPU, but it works, like, you can use it, you generate images, everything works. On Linux, on Windows, you might have a hard time, so, that's the problem, and most people, I think most people who bought AMD probably use Windows. They probably aren't going to switch to Linux, so... Yeah. So, until AMD actually, like, ports their, like, raw cam to, to Windows properly, and then there's actually PyTorch, I think they're, they're doing that, they're in the process of doing that, but, until they get it, they get a good, like, PyTorch raw cam build that works on Windows, it's, like, they're going to have a hard time. Yeah.Alessio [00:35:06]: We got to get George on it. Yeah. Well, he's trying to get Lisa Su to do it, but... Let's talk a bit about, like, the node design. So, unlike all the other text-to-image, you have a very, like, deep, so you have, like, a separate node for, like, clip and code, you have a separate node for, like, the case sampler, you have, like, all these nodes. Going back to, like, the making it easy versus making it hard, but, like, how much do people actually play with all the settings, you know? Kind of, like, how do you guide people to, like, hey, this is actually going to be very impactful versus this is maybe, like, less impactful, but we still want to expose it to you?Comfy [00:35:40]: Well, I try to... I try to expose, like, I try to expose everything or, but, yeah, at least for the, but for things, like, for example, for the samplers, like, there's, like, yeah, four different sampler nodes, which go in easiest to most advanced. So, yeah, if you go, like, the easy node, the regular sampler node, that's, you have just the basic settings. But if you use, like, the sampler advanced... If you use, like, the custom advanced node, that, that one you can actually, you'll see you have, like, different nodes.Alessio [00:36:19]: I'm looking it up now. Yeah. What are, like, the most impactful parameters that you use? So, it's, like, you know, you can have more, but, like, which ones, like, really make a difference?Comfy [00:36:30]: Yeah, they all do. They all have their own, like, they all, like, for example, yeah, steps. Usually you want steps, you want them to be as low as possible. But you want, if you're optimizing your workflow, you want to, you lower the steps until, like, the images start deteriorating too much. Because that, yeah, that's the number of steps you're running the diffusion process. So, if you want things to be faster, lower is better. But, yeah, CFG, that's more, you can kind of see that as the contrast of the image. Like, if your image looks too bursty. Then you can lower the CFG. So, yeah, CFG, that's how, yeah, that's how strongly the, like, the negative versus positive prompt. Because when you sample a diffusion model, it's basically a negative prompt. It's just, yeah, positive prediction minus negative prediction.swyx [00:37:32]: Contrastive loss. Yeah.Comfy [00:37:34]: It's positive minus negative, and the CFG does the multiplier. Yeah. Yeah. Yeah, so.Alessio [00:37:41]: What are, like, good resources to understand what the parameters do? I think most people start with automatic, and then they move over, and it's, like, snap, CFG, sampler, name, scheduler, denoise. Read it.Comfy [00:37:53]: But, honestly, well, it's more, it's something you should, like, try out yourself. I don't know, you don't necessarily need to know how it works to, like, what it does. Because even if you know, like, CFGO, it's, like, positive minus negative prompt. Yeah. So the only thing you know at CFG is if it's 1.0, then that means the negative prompt isn't applied. It also means sampling is two times faster. But, yeah. But other than that, it's more, like, you should really just see what it does to the images yourself, and you'll probably get a more intuitive understanding of what these things do.Alessio [00:38:34]: Any other nodes or things you want to shout out? Like, I know the animate diff IP adapter. Those are, like, some of the most popular ones. Yeah. What else comes to mind?Comfy [00:38:44]: Not nodes, but there's, like, what I like is when some people, sometimes they make things that use ComfyUI as their backend. Like, there's a plugin for Krita that uses ComfyUI as its backend. So you can use, like, all the models that work in Comfy in Krita. And I think I've tried it once. But I know a lot of people use it, and it's probably really nice, so.Alessio [00:39:15]: What's the craziest node that people have built, like, the most complicated?Comfy [00:39:21]: Craziest node? Like, yeah. I know some people have made, like, video games in Comfy with, like, stuff like that. So, like, someone, like, I remember, like, yeah, last, I think it was last year, someone made, like, a, like, Wolfenstein 3D in Comfy. Of course. And then one of the inputs was, oh, you can generate a texture, and then it changes the texture in the game. So you can plug it to, like, the workflow. And there's a lot of, if you look there, there's a lot of crazy things people do, so. Yeah.Alessio [00:39:59]: And now there's, like, a node register that people can use to, like, download nodes. Yeah.Comfy [00:40:04]: Like, well, there's always been the, like, the ComfyUI manager. Yeah. But we're trying to make this more, like, I don't know, official, like, with, yeah, with the node registry. Because before the node registry, the, like, okay, how did your custom node get into ComfyUI manager? That's the guy running it who, like, every day he searched GitHub for new custom nodes and added dev annually to his custom node manager. So we're trying to make it less effortless. So we're trying to make it less effortless for him, basically. Yeah.Alessio [00:40:40]: Yeah. But I was looking, I mean, there's, like, a YouTube download node. There's, like, this is almost like, you know, a data pipeline more than, like, an image generation thing at this point. It's, like, you can get data in, you can, like, apply filters to it, you can generate data out.Comfy [00:40:54]: Yeah. You can do a lot of different things. Yeah. So I'm thinking, I think what I did is I made it easy to make custom nodes. So I think that helped a lot. I think that helped a lot for, like, the ecosystem because it is very easy to just make a node. So, yeah, a bit too easy sometimes. Then we have the issue where there's a lot of custom node packs which share similar nodes. But, well, that's, yeah, something we're trying to solve by maybe bringing some of the functionality into the core. Yeah. Yeah. Yeah.Alessio [00:41:36]: And then there's, like, video. People can do video generation. Yeah.Comfy [00:41:40]: Video, that's, well, the first video model was, like, stable video diffusion, which was last, yeah, exactly last year, I think. Like, one year ago. But that wasn't a true video model. So it was...swyx [00:41:55]: It was, like, moving images? Yeah.Comfy [00:41:57]: I generated video. What I mean by that is it's, like, it's still 2D Latents. It's basically what I'm trying to do. So what they did is they took SD2, and then they added some temporal attention to it, and then trained it on videos and all. So it's kind of, like, animated, like, same idea, basically. Why I say it's not a true video model is that you still have, like, the 2D Latents. Like, a true video model, like Mochi, for example, would have 3D Latents. Mm-hmm.Alessio [00:42:32]: Which means you can, like, move through the space, basically. It's the difference. You're not just kind of, like, reorienting. Yeah.Comfy [00:42:39]: And it's also, well, it's also because you have a temporal VAE. Mm-hmm. Also, like, Mochi has a temporal VAE that compresses on, like, the temporal direction, also. So that's something you don't have with, like, yeah, animated diff and stable video diffusion. They only, like, compress spatially, not temporally. Mm-hmm. Right. So, yeah. That's why I call that, like, true video models. There's, yeah, there's actually a few of them, but the one I've implemented in comfy is Mochi, because that seems to be the best one so far. Yeah.swyx [00:43:15]: We had AJ come and speak at the stable diffusion meetup. The other open one I think I've seen is COG video. Yeah.Comfy [00:43:21]: COG video. Yeah. That one's, yeah, it also seems decent, but, yeah. Chinese, so we don't use it. No, it's fine. It's just, yeah, I could. Yeah. It's just that there's a, it's not the only one. There's also a few others, which I.swyx [00:43:36]: The rest are, like, closed source, right? Like, Cling. Yeah.Comfy [00:43:39]: Closed source, there's a bunch of them. But I mean, open. I've seen a few of them. Like, I can't remember their names, but there's COG videos, the big, the big one. Then there's also a few of them that released at the same time. There's one that released at the same time as SSD 3.5, same day, which is why I don't remember the name.swyx [00:44:02]: We should have a release schedule so we don't conflict on each of these things. Yeah.Comfy [00:44:06]: I think SD 3.5 and Mochi released on the same day. So everything else was kind of drowned, completely drowned out. So for some reason, lots of people picked that day to release their stuff.Comfy [00:44:21]: Yeah. Which is, well, shame for those. And I think Omnijet also released the same day, which also seems interesting. Yeah. Yeah.Alessio [00:44:30]: What's Comfy? So you are Comfy. And then there's like, comfy.org. I know we do a lot of things for, like, news research and those guys also have kind of like a more open source thing going on. How do you work? Like you mentioned, you mostly work on like, the core piece of it. And then what...Comfy [00:44:47]: Maybe I should fade it in because I, yeah, I feel like maybe, yeah, I only explain part of the story. Right. Yeah. Maybe I should explain the rest. So yeah. So yeah. Basically, January, that's when the first January 2023, January 16, 2023, that's when Amphi was first released to the public. Then, yeah, did a Reddit post about the area composition thing somewhere in, I don't remember exactly, maybe end of January, beginning of February. And then someone, a YouTuber, made a video about it, like Olivio, he made a video about Amphi in March 2023. I think that's when it was a real burst of attention. And by that time, I was continuing to develop it and it was getting, people were starting to use it more, which unfortunately meant that I had first written it to do like experiments, but then my time to do experiments went down. It started going down, because people were actually starting to use it then. Like, I had to, and I said, well, yeah, time to add all these features and stuff. Yeah, and then I got hired by Stability June, 2023. Then I made, basically, yeah, they hired me because they wanted the SD-XL. So I got the SD-XL working very well withітhe UI, because they were experimenting withámphi.house.com. Actually, the SDX, how the SDXL released worked is they released, for some reason, like they released the code first, but they didn't release the model checkpoint. So they released the code. And then, well, since the research was related to code, I released the code in Compute 2. And then the checkpoints were basically early access. People had to sign up and they only allowed a lot of people from edu emails. Like if you had an edu email, like they gave you access basically to the SDXL 0.9. And, well, that leaked. Right. Of course, because of course it's going to leak if you do that. Well, the only way people could easily use it was with Comfy. So, yeah, people started using. And then I fixed a few of the issues people had. So then the big 1.0 release happened. And, well, Comfy UI was the only way a lot of people could actually run it on their computers. Because it just like automatic was so like inefficient and bad that most people couldn't actually, like it just wouldn't work. Like because he did a quick implementation. So people were forced. To use Comfy UI, and that's how it became popular because people had no choice.swyx [00:47:55]: The growth hack.Comfy [00:47:56]: Yeah.swyx [00:47:56]: Yeah.Comfy [00:47:57]: Like everywhere, like people who didn't have the 4090, they had like, who had just regular GPUs, they didn't have a choice.Alessio [00:48:05]: So yeah, I got a 4070. So think of me. And so today, what's, is there like a core Comfy team or?Comfy [00:48:13]: Uh, yeah, well, right now, um, yeah, we are hiring. Okay. Actually, so right now core, like, um, the core core itself, it's, it's me. Uh, but because, uh, the reason where folks like all the focus has been mostly on the front end right now, because that's the thing that's been neglected for a long time. So, uh, so most of the focus right now is, uh, all on the front end, but we are, uh, yeah, we will soon get, uh, more people to like help me with the actual backend stuff. Yeah. So, no, I'm not going to say a hundred percent because that's why once the, once we have our V one release, which is because it'd be the package, come fee-wise with the nice interface and easy to install on windows and hopefully Mac. Uh, yeah. Yeah. Once we have that, uh, we're going to have to, lots of stuff to do on the backend side and also the front end side, but, uh.Alessio [00:49:14]: What's the release that I'm on the wait list. What's the timing?Comfy [00:49:18]: Uh, soon. Uh, soon. Yeah, I don't want to promise a release date. We do have a release date we're targeting, but I'm not sure if it's public. Yeah, and we're still going to continue doing the open source, making MPUI the best way to run stable infusion models. At least the open source side, it's going to be the best way to run models locally. But we will have a few things to make money from it, like cloud inference or that type of thing. And maybe some things for some enterprises.swyx [00:50:08]: I mean, a few questions on that. How do you feel about the other comfy startups?Comfy [00:50:11]: I mean, I think it's great. They're using your name. Yeah, well, it's better they use comfy than they use something else. Yeah, that's true. It's fine. We're going to try not to... We don't want to... We want people to use comfy. Like I said, it's better that people use comfy than something else. So as long as they use comfy, I think it helps the ecosystem. Because more people, even if they don't contribute directly, the fact that they are using comfy means that people are more likely to join the ecosystem. So, yeah.swyx [00:50:57]: And then would you ever do text?Comfy [00:50:59]: Yeah, well, you can already do text with some custom nodes. So, yeah, it's something we like. Yeah, it's something I've wanted to eventually add to core, but it's more like not a very... It's a very high priority. But because a lot of people use text for prompt enhancement and other things like that. So, yeah, it's just that my focus has always been on diffusion models. Yeah, unless some text diffusion model comes out.swyx [00:51:30]: Yeah, David Holtz is investing a lot in text diffusion.Comfy [00:51:34]: Yeah, well, if a good one comes out, then we'll probably implement it since it fits with the whole...swyx [00:51:39]: Yeah, I mean, I imagine it's going to be a close source to Midjourney. Yeah.Comfy [00:51:43]: Well, if an open one comes out, then I'll probably implement it.Alessio [00:51:54]: Cool, comfy. Thanks so much for coming on. This was fun. Bye. Get full access to Latent Space at www.latent.space/subscribe

The Agents of Change: SEO, Social Media, and Mobile Marketing for Small Business

What if you could turn your creative visions into reality without hiring an expensive designer? That's where AI tools like MidJourney and Ideogram come in. On this episode of The Agents of Change, I sit down with Jeff Sieh, a content creator, live streamer, and AI enthusiast, to explore how marketers and creators can use AI to craft stunning visuals and repurpose content more efficiently. Jeff's insights into prompt engineering, style references, and blending tools will help you cut through the noise and stand out in a crowded feed. If you've ever struggled with getting the perfect image or want to better leverage AI in your marketing efforts, this episode is for you. https://www.theagentsofchange.com/562

TeknoSafari's Podcast
Çakal Stajyer Yapay Zeka Ajanlığı Yaparsa - Yapay Zekada Bu Hafta S2 B4

TeknoSafari's Podcast

Play Episode Listen Later Nov 14, 2024 19:57


1. Claude bilgisayarınızı görebiliyor! 2. Perplexity NotebookLM'e rakip olmaya çalışıyor. Ayrıca Reasoningle de iddialı. Öte yandan #WallStreetJournal ve #NewYorkPost telif hakkı ihlali ve marka hasarıyla suçlayarak #PerplexityAI ye dava açıyor. 3. Ideogram, CANVAS ile zirveye çıktı. 4. SOTA videoda çok iyi ve açık kaynaklı. Mochi 1 kesinlikle denemeye değer. 5. Flux gelince konu kapanmadı, Stabble Diffusion 3.5 geldi 6. RunwayML, At-One ile atakta 7. OpenAI gelişmiş sesi Avrupa'ya da açtı. 8. ByteDance stajyeri yapay zeka modellerine zararlı kod yerleştirdiği için kovuldu 9. Elon XAI APIsi yayınladı. Grok uygulamalarınıza eklenebiliyor. GROK3 gelirse ortalık karışır. 10. GPT Windows uygulaması geldi. #yapayzeka #teknolojihaberleri #bilim

OMT - Webinare
Marketing neu denken: Wie du mit KI im Marketing die Nase vorn hast (Timo Brümmer & Leon Peters)

OMT - Webinare

Play Episode Listen Later Nov 6, 2024 67:07


Erfahre in diesem Webinar, wie du durch den gezielten Einsatz von Künstlicher Intelligenz das volle Potenzial deines Marketings ausschöpfst. Von der Effizienzsteigerung von Prozessen bis hin zur personalisierten Kundenansprache – wir zeigen dir praxisnah, wie KI dein Marketing erfolgreicher macht. Dabei stellen wir konkrete Anwendungsfälle mit den Tools ChatGPT, Ideogram.ai, Midjourney, Runwayml.com sowie Udio.com vor. Folgendes lernst Du in diesem Webinar: – Du erhältst eine Einführung in KI und welche Bedeutung sie für deinen Arbeitsalltag hat – Du bekommst konkrete Praxisbeispiele für Text, Bild, Video und Audio, um KI in deine Arbeit zu integrieren – Du lernst die Chancen und Grenzen der KI-Tools kennen – Du bekommst einen exklusiven Tipp, wie du dein KI-Wissen auf dem aktuellen Stand hältst

Let's Talk AI
#187 - Anthropic Agents, Mochi1, 3.4B data center, OpenAI's FAST image gen

Let's Talk AI

Play Episode Listen Later Oct 28, 2024 129:38


Our 187th episode with a summary and discussion of last week's big AI news, now with Jeremie co-hosting once again! With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris) Read out our text newsletter and comment on the podcast at https://lastweekin.ai/. If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form. Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai Timestamps + Links: (00:00:00) Intro / Banter (00:03:07) Response to listener comments / corrections (00:05:13) Sponsor Read) Tools & Apps(00:06:22) Anthropic's latest AI update can use a computer on its own (00:18:09) AI video startup Genmo launches Mochi 1, an open source rival to Runway, Kling, and others (00:20:37) Canva has a shiny new text-to-image generator (00:23:35) Canvas Beta brings Remix, Extend, and Magic Fill to Ideogram users (00:26:16) StabilityAI releases Stable Diffusion 3.5  (00:28:27) Bringing Agentic Workflows into Inflection for Enterprise Applications & Business(00:32:35) Crusoe's $3.4B joint venture to build AI data center campus with up to 100,000 GPUs (00:39:08) Anthropic reportedly in early talks to raise new funding on up to $40B valuation (00:45:47) Longtime policy researcher Miles Brundage leaves OpenAI (00:49:53) NVIDIA's Blackwell GB200 AI Servers Ready For Mass Deployment In December (00:52:41) Foxconn building Nvidia superchip facility in Mexico, executives say (00:55:27) xAI, Elon Musk's AI startup, launches an API Projects & Open Source(00:58:32) INTELLECT-1: The First Decentralized 10-Billion-Parameter AI Model Training (01:06:34) Meta FAIR Releases Eight New AI Research Artifacts—Models, Datasets, and Tools to Inspire the AI Community (01:10:02) Google DeepMind is making its AI text watermark open source Research & Advancements(01:13:21) OpenAI researchers develop new model that speeds up media generation by 50X (01:17:54) How much AI compute is out there, and who owns it? (01:25:28) Rewarding Progress: Scaling Automated Process Verifiers for LLM Reasoning (01:33:30) Inference Scaling for Long-Context Retrieval Augmented Generation Policy & Safety(01:41:50) Announcing our updated Responsible Scaling Policy (01:48:52) Anthropic is testing AI's capacity for sabotage (01:56:30) OpenAI asked US to approve energy-guzzling 5GW data centers, report says (02:00:05) US Probes TSMC's Dealings with Huawei (02:03:03) TikTok owner ByteDance taps TSMC to make its own AI GPUs to stop relying on Nvidia — the company has reportedly spent over $2 billion on Nvidia AI GPUs (02:06:37) Outro

The top AI news from the past week, every ThursdAI

Hey all, Alex here, coming to you from the (surprisingly) sunny Seattle, with just a mind-boggling week of releases. Really, just on Tuesday there was so much news already! I had to post a recap thread, something I do usually after I finish ThursdAI! From Anthropic reclaiming close-second sometimes-first AI lab position + giving Claude the wheel in the form of computer use powers, to more than 3 AI video generation updates with open source ones, to Apple updating Apple Intelligence beta, it's honestly been very hard to keep up, and again, this is literally part of my job! But once again I'm glad that we were able to cover this in ~2hrs, including multiple interviews with returning co-hosts ( Simon Willison came back, Killian came back) so definitely if you're only a reader at this point, listen to the show! Ok as always (recently) the TL;DR and show notes at the bottom (I'm trying to get you to scroll through ha, is it working?) so grab a bucket of popcorn, let's dive in

AI For Humans
Anthropic's New AI Agent, OpenAI Plays Catch-up, Runway's Act-One & More AI News

AI For Humans

Play Episode Listen Later Oct 24, 2024 50:12


AI NEWS: Agents are here from Anthropic with Computer Use in Claude Sonnet 3.5 (new) and likely coming from OpenAI, O1 keeps getting better and might get upgraded soon, Runway's New Act One let's you puppet AI video, Ideogram's new Canvas upgrades AI imaging, Unitree's Robots are getting WAY better and we show you how to make Google's NotebookLM uncensored. AND OH SO MUCH MORE.   It's a big, massive week of AI news. And we are here, for you.   Join our Patreon: https://www.patreon.com/AIForHumansShow Jump in our Discord: https://discord.gg/muD2TYgC8f Follow us for more on X @AIForHumansShow Join our TikTok @aiforhumansshow And to contact or book us for speaking/consultation, please visit our website: https://www.aiforhumans.show/   // Show Links //   Anthropic Drops “Computer Use” In Sonnet 3.5 aka AI Agents https://www.anthropic.com/news/3-5-models-and-computer-use   Claude Coding 90s Website: https://youtu.be/vH2f7cjXjKI?si=XqTRKVxHZx1bK36b   Picks the first link on Google: https://x.com/AnthropicAI/status/1848742757151498717   What Computer Use Can't Do https://x.com/forgebitz/status/1848764235729244254   OpenAI's Noam Brown on O1 https://v.redd.it/7dic62adm3wd1   OpenAI Feels The Pressure, Close To Releasing Coding Bot https://www.theinformation.com/articles/openai-in-duel-with-anthropic-doubles-down-on-ai-that-writes-software   OpenAI Agentic Rumors Involving Microsoft https://x.com/flowersslop/status/1848506100435304852   Sam Altman Teases ChatGPT Update For Second Birthday https://x.com/sama/status/1848487309211275398   Satya Nadella Says We're “Using AI Tools to Build Better AI” https://x.com/tsarnick/status/1848472478257189374   Runway Act-One https://runwayml.com/research/introducing-act-one   Teaser Video https://x.com/runwayml/status/1848785907723473001   Two actors in a scene https://x.com/runwayml/status/1848785913918218517   Mochi 1 -- New OpenSource AI Video From Genmo https://x.com/genmoai/status/1848762405779574990   Ideogram Canvas Feature https://x.com/ideogram_ai/status/1848757699606983143   Stable Diffusion 3.5 https://x.com/StabilityAI/status/1848729212250951911   Unitree Robot Exercise Videos https://youtu.be/G6JE7mNYz2A?si=KLiXYznOUy7Qz4Rh   TANGO https://x.com/dreamingtulpa/status/1847310594434584922   Trump at a McDonald's https://x.com/aliensupershow/status/1848438728148111822   NotebookLM Uncensored https://www.reddit.com/r/notebooklm/comments/1g64iyi/holy_shit_listeners_notebooklm_can_generate_18/

Raffaele Gaito, il podcast.

Ormai abbiamo a disposizione tantissimi strumenti di AI per generare immagini.Midjourney, Leonardo, Ideogram, Dalle e tanti altri...Come orientarsi in questa vasta scelta di alternative? Qual è la migliore? A quale conviene abbonarsi a pagamento?Ecco il confronto che stavate aspettando dove vi suggerisco lo strumento text-to-image più adatto alle vostre esigenze... e anche alle vostre tasche.Buon ascolto

da Brand a Friend
#341 - Di Chi Fidarsi

da Brand a Friend

Play Episode Listen Later Oct 13, 2024 21:32


Quando sei in un posto nuovo, con gente, lingua, cultura e tradizioni diverse da quelle che conosci è importante essere in grado di riconoscere e valutare rapidamente chi sono le persone di cui ti puoi fidare.Nonostante la maggior parte delle persone come te pensino che questo sia un compito più che semplice e ampliamente alla loro portata, quello che osservo è che non è affatto così.Tendiamo a fidarci di persone più per le emozioni che ci trasmettono così come per motivi di simpatia, similarità, e a seguito di atti di cortesia e generosità, quando dovremmo invece prestare attenzione ad elementi a volte meno evidenti, ma assai più importanti nel valutare l'affidabilità, credibilità e la fiducia da concedere ad una persona.Non è facile, ma con il tempo e l'esperienza sviluppi dei sensori speciali e una capacità intuitiva che ti aiuta a valutare rapidamente l'affidabilità di una persona nuova.Questi - qui di seguito - sono i criteri personali che uso per capire, il più velocemente possibile, se una persona è affidabile in un contesto lavorativo.Li ho imparati a causa di molti errori e perché ho dato fiducia a troppe persone che non la meritavano.Vorrei condividere questi criteri per offrire una prospettiva personale, non accademica, su cosa osservare per capire di chi fidarti quando sei lontano da casa, e per aiutarti a diventare più consapevole di certi comportamenti che potresti voler rivedere.Comportamenti delle Persone di cui Non Dovremmo Fidarci1) Non assumersi responsabilità - Dare la colpa ad altroGiustificarsi  Non prendersi la responsabilità per ciò che va storto – dare la colpa agli altri, agli eventi, all'imprevedibilità delle cose. Accusare.Comportamento di alta fiducia:  Lodare gli altri.  Assumersi la piena responsabilità per tutto.2) Ego-CentrismoVantarsi / Esaltare - Evangelizzazione Non Richiesta  “Noi facciamo questo, noi facciamo quello...”. “Sarà un successo.” “Siamo i migliori.”Voler impressionare gli altri. Insistenza nel persuadere ed esporre agli altri (non richiesto) come funzionano le cose.Comportamento di alta fiducia:  Tenere un profilo basso.  Essere molto umili riguardo ai successi futuri.3) Sicurezza in Sé Stessi  Minimizzare i rischi – atteggiamento di “non c'è problema”.“Non ti preoccupare...” come mantra ufficiale.Comportamento di alta fiducia:  Identificare tutti i possibili rischi.  Avere un piano B per ogni rischio.  Pianificare per il peggio.  Prepararsi agli eventi imprevisti, ai problemi e agli incidenti.4) Improvvisare  Fare piani all'ultimo minuto – salvare le situazioni per un pelo – agire all'ultimo secondo. Poco o nessun piano. Orari degli appuntamenti vaghi.Comportamento di alta fiducia:  Pianificazione approfondita.  Anticipare i problemi.  Orari precisi per gli appuntamenti.5) Parlare-Parlare  Concentrarsi più sul parlare che sull'ascoltare. Bassa capacità di ascoltare attentamente e rispondere agli input. Focalizzato sull'esporre le proprie idee e opinioni.Comportamento di alta fiducia:  Ascoltare attentamente.  Fare domande pertinenti.  Prendere appunti.  Non interrompere.6) Frequentazioni SbagliatePassare del tempo con persone di bassa fiducia.Comportamento di alta fiducia:  Frequentare persone di alta fiducia.  Ridurre il tempo passato con persone non affidabili.  Cercare e coltivare relazioni con persone di alta fiducia.7) Parlare Male di Chi Non C'è  Sparlare di altri, collaboratori o concorrenti – criticare e dare la colpa ad altri che non sono presenti.Comportamento di alta fiducia:  Lodare gli altri.  Adottare un atteggiamento di “imparare dagli altri”.  Usare silenziosamente gli errori altrui come opportunità di business.8) Arrivare in RitardoNon Essere Puntuali  Essere in ritardo agli appuntamenti.Comportamento di alta fiducia:  Essere sempre puntuali o in anticipo.9) Sempre di Corsa - con Poco TempoFare Tutto di Fretta  Essere stressati per troppe cose da fare. Avere poco tempo – sempre di corsa. Poco o nessun delegare. Fare le cose con poco tempo. Tagliare gli angoli e agire in modo approssimativo per mancanza di tempo. Prendere decisioni in fretta, sotto pressione.Comportamento di alta fiducia:  Buona pianificazione giornaliera.  Sapere come dare priorità.  Riservare abbastanza tempo per le attività importanti.  Ridurre drasticamente il numero di cose importanti da fare in un giorno.  Delegare.10) Interrompere  Intervenire quando gli altri parlano senza lasciarli finire. Aggiungere, interrompere e rilanciare continuamente quello che gli altri stanno dicendo senza aspettare che finiscano.Comportamento di alta fiducia:  Ascoltare attentamente.  Prendere appunti.  Riflettere su ciò che si è compreso.11) Saltare da un Argomento all'Altro  Passare da un argomento all'altro senza mai completarne uno e senza arrivare a una conclusione.Comportamento di alta fiducia:  Presentazione ordinata.  Analizzare un argomento alla volta.  Arrivare a conclusioni/decisioni prima di cambiare argomento.12) Discutere  Adottare un approccio di discussione durante le riunioni di lavoro.Comportamento di alta fiducia:  Stabilire gli obiettivi da raggiungere durante la riunione.  Ascoltare e tenere in considerazione tutti i punti di vista.  Arrivare ad accordi prima di andare avanti.13) Mancanza di Trasparenza  Nascondere conflitti di interesse. Minimizzare o nascondere fallimenti personali, problemi o questioni.Comportamento di alta fiducia:  Riconoscere e ammettere pubblicamente i fallimenti personali.  Scusarsi.  Agire proattivamente per correggere gli errori._______________Info Utili• Sostieni questo podcast:Ottieni feedback, ricevi consigli sul tuo progetto onlinehttps://Patreon.com/Robin_Good•  Musica di questa puntata:"Ergo" by Birocratic disponibile su Bandcamp•  Nella foto di copertina:Dogs sniffing each other. Image generated with Ideogram.ai• Dammi feedback:Critiche, commenti, suggerimenti, idee e domande unendoti al gruppo Telegram https://t.me/@RobinGoodPodcastFeedback• Ascolta e condividi questo podcast:https://www.spreaker.com/show/dabrandafriendArchivio completo organizzato per temi:https://start.me/p/kxENzk/da-brand-a-friend-archivio-podcast• Seguimi su Telegram:https://t.me/RobinGoodItalia• Newsletter in Inglese:https://robingood.substack.comhttps://goodtools.substack.comhttps://curationmonetized.substack.com 

Marketing Against The Grain
NotebookLM is INSANE! How to Use Google's AI Tool for Marketing In 2024

Marketing Against The Grain

Play Episode Listen Later Oct 10, 2024 18:24


Ep. 268 "Is AI about to revolutionize storytelling forever?" Kipp dives into the groundbreaking potential of Google's NotebookLM and how it might reshape marketing strategies in 2024. Learn more about leveraging AI for creating engaging media, how engineers are becoming pivotal in marketing, and tips on automating content creation efficiently. Mentions Andrej Karpathy https://x.com/karpathy Histories of Mysteries podcast https://open.spotify.com/show/3K4LRyMCP44kBbiOziwJjb?si=432a337c28f14d97&nd=1&dlsi=1e2cd9b320094415 NotebookLM https://notebooklm.google/ Ideogram https://ideogram.ai/ Claude https://claude.ai/ Nathan Barry https://x.com/nathanbarry Resource [Free] Steal our favorite AI Prompts featured on the show! Grab them here: https://clickhubspot.com/aip We're on Social Media! Follow us for everyday marketing wisdom straight to your feed YouTube: ​​https://www.youtube.com/channel/UCGtXqPiNV8YC0GMUzY-EUFg  Twitter: https://twitter.com/matgpod  TikTok: https://www.tiktok.com/@matgpod  Join our community https://landing.connect.com/matg Thank you for tuning into Marketing Against The Grain! Don't forget to hit subscribe and follow us on Apple Podcasts (so you never miss an episode)! https://podcasts.apple.com/us/podcast/marketing-against-the-grain/id1616700934   If you love this show, please leave us a 5-Star Review https://link.chtbl.com/h9_sjBKH and share your favorite episodes with friends. We really appreciate your support. Host Links: Kipp Bodnar, https://twitter.com/kippbodnar   Kieran Flanagan, https://twitter.com/searchbrat  ‘Marketing Against The Grain' is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Produced by Darren Clarke.

Prudent Observations with The Prudentialist
The Discourse Corner #2: The Ideogram of a Smoking Gun

Prudent Observations with The Prudentialist

Play Episode Listen Later Oct 1, 2024 52:35


Friend of the show Capitalisimo of The Sidebar Podcast joins me to discuss the history of the gun emoji, "weaponization" of communication in the culture war, with references to Jeff Deist, John Waters, and more. This is the free hour, with the full show available on Subscribestar and Substack. See The Full Show: https://www.subscribestar.com/the-prudentialist Merch: https://mr-prudes-wares.creator-spring.com/ All other links: https://findmyfrens.net/theprudentialist/

DigitalFeeling
Episode 89 - IA et Design, quels sont les outils à utiliser ?

DigitalFeeling

Play Episode Listen Later Sep 27, 2024 14:46


L'IA Générative et le design créatif : Révolutionner la création visuelle avec DALL-E, Midjourney et autres Outils d'IADans cet épisode, je plonge dans le monde de l'intelligence artificielle générative et son impact sur le design créatif. L'IA générative, bien au-delà des applications classiques de l'IA comme les voitures autonomes ou les robots, se concentre ici sur des outils qui améliorent notre productivité et stimulent notre créativité artistique, tels que DALL-E, Midjourney, et d'autres.Qu'est-ce que l'IA Générative ?L'IA générative utilise des algorithmes avancés pour créer des contenus originaux en combinant des données existantes de manière novatrice. Contrairement à la simple reproduction, elle est capable de générer des visuels uniques à partir de descriptions textuelles ou d'autres inputs, permettant aux designers et artistes de repousser les limites de leur imagination sans les contraintes de temps ou de ressources.Outils et technologies d'IA GénérativeDALL-E : Développé par OpenAI, cet outil utilise des modèles de langage pour générer des images à partir de descriptions textuelles. Accessible via des plateformes comme ChatGPT ou Canva, DALL-E est particulièrement utile pour les designers souhaitant créer des visuels sur mesure de manière intuitive.Midjourney : Très apprécié pour sa capacité à produire des images artistiques avec un style unique, Midjourney est désormais accessible via une application web simplifiée. Idéal pour des concepts de branding ou de storytelling visuel, cet outil demande un certain apprentissage pour maîtriser les "prompts", mais offre des résultats souvent époustouflants.Fluux : Un nouvel outil qui se distingue par sa simplicité d'utilisation et la qualité de ses rendus. Fluux permet de générer des images en fonction de prompts moins complexes, tout en offrant une version freemium pour les utilisateurs.Ideogram : Spécialement conçu pour intégrer du texte dans les images, Ideogram se distingue par sa gestion efficace des créations visuelles qui incluent des éléments textuels, une fonctionnalité souvent problématique pour d'autres outils d'IA.Applications en design créatifL'IA générative transforme le workflow des designers en accélérant le processus créatif, facilitant la génération rapide de prototypes et l'exploration de multiples concepts artistiques. Par exemple, Nike utilise l'IA pour créer des designs de chaussures personnalisées basés sur les préférences des utilisateurs et des tendances actuelles, tandis que Coca-Cola génère des étiquettes personnalisées inspirées par des préférences culturelles et des événements actuels.Avantages et défis de l'IA GénérativeAvantages :Créativité amplifiée : L'IA offre une source inépuisable d'inspiration, permettant aux designers de découvrir des combinaisons visuelles inédites et d'aller au-delà des idées traditionnelles.Personnalisation de masse : En adaptant les créations aux préférences individuelles, les marques peuvent augmenter la pertinence et l'impact de leurs campagnes marketing.Gain de temps et de ressources : En automatisant une partie du processus créatif, les designers peuvent se concentrer sur l'affinement de leur vision artistique.Défis :Contrôle de la qualité : Les résultats peuvent parfois manquer de cohérence ou ne pas atteindre les standards de qualité requis, nécessitant plusieurs itérations de prompts.Cohérence de marque : Intégrer des créations générées par l'IA dans une stratégie de marque peut poser des défis en termes de maintien de l'identité visuelle et de la cohérence des messages.L'avenir de l'IA Générative dans le DesignÀ mesure que les technologies d'IA générative continuent de s'améliorer, elles joueront un rôle croissant dans le design créatif, offrant aux marques des opportunités pour se démarquer par des visuels uniques et captivants. Cependant, il est essentiel de réfléchir à la valeur ajoutée de l'IA dans chaque campagne et de s'assurer que son utilisation est en accord avec la stratégie de marque pour éviter les bad buzz, comme cela a pu être le cas avec l'office de tourisme de Chamonix.Écoutez cet épisode et découvrez comment l'IA générative peut transformer votre approche du design et offrir des expériences visuelles innovantes à vos clients !Vous avez aimez cet épisode et vous voulez soutenir ce podcast ? Laissez moi un avis sur Apple podcast ou Spotify et parlez-en autour de vous, il pourrait aider quelqu'un d'autre ;)Vous avez une question en marketing digital ou sur l'Intelligence Artificielle Générative ? Posez-moi vos questions sur Linkedin en MP et je me ferai un plaisir d'y répondre lors d'un prochain épisode : https://www.linkedin.com/in/elodiechenolPour rester en contact découvrez ma newsletter : https://substack.com/@elodiechenolA bientôt pour un prochain épisode ;)Et si vous me découvrez à travers ce podcast, je suis Elodie Chenol, formatrice, consultante et conférencière en marketing digital et Intelligence artificielle.A bientôt !Elo

The Next Wave - Your Chief A.I. Officer
5+ AI Workflows You Can Copy For Your Business in 2024

The Next Wave - Your Chief A.I. Officer

Play Episode Listen Later Sep 10, 2024 43:08


Episode 23: How can AI simplify complex workflows and enrich language learning? Matt Wolfe (https://x.com/mreflow)) and Nathan Lands (https://x.com/NathanLands) take you on an insightful journey exploring diverse AI use cases and tools. In this episode, Matt and Nathan delve into the intricacies of learning Japanese and coding with AI, leveraging tools like Claude and Perplexity to streamline and enhance these processes. Nathan shares his experience simplifying Japanese language learning through targeted translation techniques, while Matt reveals his tips for efficient coding using AI, along with strategies for optimizing content with AI tools like Perplexity and Ideogram. The duo also discusses workflow automation, potential SEO hacks, and meeting management with AI, rounding out the episode with engaging and valuable insights. Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd — Show Notes: (00:00) Exploring advanced AI use cases for businesses. (04:40) AI recreates Doom from gameplay videos accurately. (07:30) Useful for newsletters, business documents, and data management. (09:55) Custom instructions enhance model quality and results. (13:09) Create shorthands for tasks using custom instructions. (18:19) Perplexity's page feature generates mini Wikipedia entries. (25:25) Curious about AI-generated YouTube thumbnail prompts. (26:48) Automating business workflows. (31:49) Automated workflows simplify data analysis with make.com. (35:31) AI efficiently simplifies meeting agenda creation process. (37:21) Google Meet now summarizes and generates meeting notes. — Mentions: Riley Brown: https://www.youtube.com/channel/UCMcoud_ZW7cfxeIugBflSBw Ideogram 2.0: https://about.ideogram.ai/2.0 Perplexity: https://www.perplexity.ai/ Claude: https://claude.ai/ Make.com: https://www.make.com/en — Check Out Matt's Stuff: • Future Tools - https://futuretools.beehiiv.com/ • Blog - https://www.mattwolfe.com/ • YouTube- https://www.youtube.com/@mreflow — Check Out Nathan's Stuff: Newsletter: https://news.lore.com/ Blog - https://lore.com/ The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

The Family History AI Show
EP12: Hollywood AI Blunder, AI Image Generator Roundup, Google Lens Saves You Time Researching, Use AI For Translation

The Family History AI Show

Play Episode Listen Later Sep 10, 2024 57:09


Hosts Mark Thompson and Steve Little expertly navigate the rapidly evolving AI landscape, offering practical insights on leveraging AI for your family history research. Mark and Steve open this episode by discussing lessons that genealogists can learn from Hollywood on the importance of fact-checking. Next, they provide a comprehensive roundup of the top AI image-generation tools. Then, how Google Lens' big upgrade can greatly simplify your research. In this week's Tip of the Week, learn how AI translation can help you research in another language, and so much more. With a mix of news, analysis, and hands-on advice, this podcast equips you with the knowledge to harness AI's power in uncovering your family stories. Whether you're a tech-savvy researcher or new to AI, this show offers valuable insights.TimestampsI. AI In the News00:01:16 Hollywood AI Mistake: Lessons for genealogists on fact-checking.00:04:18 AI Image Generators: Overview of popular tools and their applications.00:24:22 Google Lens Upgrade: New features for image and text search.II. Tip of the Week00:30:13 AI Building Blocks - Translation: Exploring translation and applications in genealogy.III. RapidFire00:41:57 Microsoft Edge Tab Organizer: New AI-powered feature for efficient research.00:44:18 "The AI Scientist": Discussion of AI agents and complex problem-solving.00:49:10 Eleven Labs Reader Upgrade: Text-to-speech tool now supports 32 languages.00:53:03 OpenAI's Condé Nast Deal: Improves AI training and new ways to access publications.Resource LinksChatGPT: https://chat.openai.com/DALL-E: https://openai.com/dall-e-2MidJourney: https://www.midjourney.com/Google Lens: https://lens.google.com/Imagen: https://deepmind.google/technologies/imagen-3/Ideogram: https://ideogram.ai/Adobe Firefly: https://www.adobe.com/sensei/generative-ai/firefly.htmlEleven Labs Reader: https://elevenlabs.io/Google Translate: https://translate.google.com/Microsoft Edge (for Tab Organizer): https://www.microsoft.com/edgeTranskribus: https://readcoop.eu/transkribus/OpenAI: https://openai.com/TagsArtificial Intelligence, Genealogy, Family History, AI Tools, Image Generation, Google Lens, OCR Technology, Language Translation, AI Ethics, Research Techniques, DALL-E, Midjourney, Adobe Firefly, Google Imagen, Microsoft Edge, Tab Management, AI Agents, Text-to-Speech, Eleven Labs Reader, OpenAI Partnerships

Inteligencia Artificial
Nuevas Herramientas IA Ideogram, Grok 2 y Cerebras

Inteligencia Artificial

Play Episode Listen Later Sep 4, 2024


En el episodio más reciente del podcast «Inteligencia Artificial», analizamos las últimas innovaciones en el mundo de la IA. Hoy exploramos tres herramientas clave que están transformando el campo de la inteligencia artificial: Ideogram v2, Grok 2, y Cerebras. Si estás interesado en el presente y futuro de la IA, sigue leyendo para conocer las […] Origen

Let's Talk AI
#180 - Ideogram v2, Imagen 3, AI in 2030, Agent Q, SB 1047

Let's Talk AI

Play Episode Listen Later Sep 3, 2024 125:23 Transcription Available


Our 180th episode with a summary and discussion of last week's big AI news! With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris) If you would like to get a sneak peek and help test Andrey's generative AI application, go to Astrocade.com to join the waitlist and the discord. Read out our text newsletter and comment on the podcast at https://lastweekin.ai/ If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form. Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai Episode Highlights: Ideogram AI's new features, Google's Imagine 3, Dream Machine 1.5, and Runway's Gen3 Alpha Turbo model advancements. Perplexity's integration of Flux image generation models and code interpreter updates for enhanced search results.  Exploration of the feasibility and investment needed for scaling advanced AI models like GPT-4 and Agent Q architecture enhancements. Analysis of California's AI regulation bill SB1047 and legal issues related to synthetic media, copyright, and online personhood credentials. Timestamps + Links: (00:00:00) Intro / Banter (00:01:08) Response to Listener Comments / Corrections Tools & Apps (00:03:58) Ideogram AI expands its features with v2 model and color palette options (00:07:48) Google Releases Powerful AI Image Generator You Can Use for Free (00:11:41) Perplexity adds Flux.1 model for Pro users alongside Playground v3 update (00:13:58) Luma drops Dream Machine 1.5 — here's what's new (00:17:49) Runway's Gen-3 Alpha Turbo is here and can make AI videos faster than you can type (00:20:21) Perplexity's latest update improves code interpreter, charts included Applications & Business (00:24:14) AMD buying server maker ZT Systems for $4.9 billion as chipmakers strengthen AI capabilities (00:28:55) Ars Technica content is now available in OpenAI services (00:34:08) Anysphere, a GitHub Copilot rival, has raised $60M Series A at  $400M valuation from a16z, Thrive, sources say 00:38:32 Stability AI appoints new Chief Technology Officer (00:41:45) Cruise's robotaxis are coming to the Uber app in 2025 Projects & Open Source (00:44:16) AI21 Introduces the Jamba Model Family: The most powerful and efficient long-context models for the enterprise (00:53:47) Microsoft reveals Phi-3.5 — this new small AI model outperforms Gemini and GPT-4o (00:57:33) Nvidia's Llama-3.1-Minitron 4B is a small language model that punches above its weight (01:00:58) Open source Dracarys models ignite generative AI fired coding Research & Advancements (01:12:35) Can AI Scaling Continue Through 2030? (01:15:35) Agent Q: Advanced Reasoning and Learning for Autonomous AI Agents (01:23:58) Transformers to SSMs: Distilling Quadratic Knowledge to Subquadratic Models (01:31:18) Loss of plasticity in deep continual learning Policy & Safety (01:38:20) California weakens bill to prevent AI disasters before final vote, taking advice from Anthropic (01:48:14) Personhood credentials: Artificial intelligence and the value of privacy-preserving tools to distinguish who is real online (01:52:44) Showing SAE Latents Are Not Atomic Using Meta-SAEs Synthetic Media & Art (01:58:33) Authors sue Claude AI chatbot creator Anthropic for copyright infringement (01:59:32) Artists' lawsuit against Stability AI and Midjourney gets more punch (02:01:43) Outro

The AI Breakdown: Daily Artificial Intelligence News and Discussions
The 5 Most Important Stories in AI This Week

The AI Breakdown: Daily Artificial Intelligence News and Discussions

Play Episode Listen Later Aug 24, 2024 12:33


Covering the five most significant stories in AI this week. Major product releases like MidJourney's new web-based interface and Ideogram 2.0's text generation capabilities are highlighted, along with key enterprise AI developments from Salesforce. The ongoing debate over California's AI safety bill SB 1047 is also discussed, as well as how intense competition among AI companies is driving benefits for consumers and startups. Stay updated on the latest in AI! Concerned about being spied on? Tired of censored responses? AI Daily Brief listeners receive a 20% discount on Venice Pro. Visit ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠https://venice.ai/nlw ⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠⁠and enter the discount code NLWDAILYBRIEF. Learn how to use AI with the world's biggest library of fun and useful tutorials: https://besuper.ai/ Use code 'podcast' for 50% off your first month. The AI Daily Brief helps you understand the most important news and discussions in AI. Subscribe to the podcast version of The AI Daily Brief wherever you listen: https://pod.link/1680633614 Subscribe to the newsletter: https://aidailybrief.beehiiv.com/ Join our Discord: https://bit.ly/aibreakdown

AIA Podcast
Новинки Google, Apple и xAI / Загруженный интеллект, AGI, любовь и роботы / AIA Podcast #40

AIA Podcast

Play Episode Listen Later Aug 24, 2024 178:16


This Day in AI Podcast
EP74: Human Eggs with Ideogram 2.0, Phi 3.5 Boom Factor + AI-Free Startups

This Day in AI Podcast

Play Episode Listen Later Aug 23, 2024 73:09


Sign up to Simtheory for an AI workspace: https://simtheory.aiTry ideogram 2.0 on Simtheory---CHAPTERS:00:00 - Ideogram 2.0: Your new AI graphics designer?23:46 - Microsoft Phi 3.5 Initial Impressions & Thoughts + Boom Factor38:51 - AI workspace productivity: how much is your productivity worth?55:08 - Procreate's Anti AI Movement: Marketing or a New Category?1:07:06 - Chris's thoughts on Phi-3.5 Fine Tuning & Lack of Documentation, Accessibility of Models to Try---To see images from the show join our Discord community: https://thisdayinai.comShow notes: https://thisdayinai.com/bookmarks/68-ep74Thanks for listening, your comments, reviews and support of the show. We really appreciate it and love hearing from you.PS. Tasmanian YouTuber Chris mentions: https://www.youtube.com/@UCalOFVbIxEAWIV5LHGkKcnw

Everyday AI Podcast – An AI and ChatGPT Podcast
EP 342: New AI Image Generators You Won't Believe - Flux, Ideogram 2, and more

Everyday AI Podcast – An AI and ChatGPT Podcast

Play Episode Listen Later Aug 22, 2024 54:13


Send Everyday AI and Jordan a text messageWin a free year of ChatGPT or other prizes! Find out how.Y'all won't believe these new AI image generators.... They're starting to blur the line between fact and fiction. (In both a good and a bad way.) And in the past few weeks, the entire game has changed. We go over what's new and what you need to know. Newsletter: Sign up for our free daily newsletterMore on this Episode: Episode PageJoin the discussion: Ask Jordan questions on AI image generatorsRelated Episodes: Ep 218: Winning the Probability Game in AI VisualsEp 198: Midjourney V6 – What's new and producing powerful ad creativesUpcoming Episodes: Check out the upcoming Everyday AI Livestream lineupWebsite: YourEverydayAI.comEmail The Show: info@youreverydayai.comConnect with Jordan on LinkedInTopics Covered in This Episode:1. Different image generators and their uses2. Importance of AI images3. AI image generation timeline4. Potential misuse of AI image generators5. AI image generation promptsTimestamps:02:50 Daily AI news06:20 AI image vs real image test09:12 AI creates images from text descriptions using AI.10:05 AI image generators use diffusion models, face copyright concerns.13:43 DALL-E and OpenAI in AI image game.18:14 Live demonstration, prompt, show, vote for best.21:30 Open source model used for image generators.27:49 Issues showing small URLs, presenting photo options.30:46 Interface is a love-hate experience with AI.34:47 Generate, observe real-time photo realistic beach image.38:25 Imprinting painting with quick color changes attempt.41:57 Results of AI image generator with social platform.42:41 Explore AI image generators for various industries.48:36 Former president spreads problematic AI-generated images.51:16 AI image generators creating videos are powerful.Keywords:AI tools, photorealistic images, ideogram, Imagine 3, Flux 1, image generation, Grok, Google DeepMind, Jordan Wilson, Midjourney, user control, AI image generators, misinformation and disinformation, open source models, copyright concerns, DALL E, diffusion models, AI-generated images, AI demonstration, Chat GPT course, webinar, audience vote, AI impact on creativity and business, Google and California deal, Microsoft AI feature recall, Neuralink brain chip, misuse of image generators, video creation, ad campaigns, newsletter giveaway Get more out of ChatGPT by learning our PPP method in this live, interactive and free training! Sign up now: https://youreverydayai.com/ppp-registration/

AI For Humans
Ideogram & Flux Make AI Imaging Too Realistic, OpenAI Updates & More AI News

AI For Humans

Play Episode Listen Later Aug 22, 2024 47:55


Join our Patreon: https://www.patreon.com/AIForHumansShow AI news coming in HOT: Ideogram 2.0 dropped and, along with Flux community updates & Google's IMAGEN-3, continues to show how AI imaging tools keep improving… but also opens the door to whole new messes. Plus, OpenAI's drops new blog posts (wee!), Unitree sends their cheap humanoid robot into production and we use Hedra and a bunch of other AI tools to interview Baby Joe Brogan. It's quite a moment.    Follow us for more on X @AIForHumansShow Join our TikTok @aiforhumansshow And to contact or book us for speaking/consultation, please visit our website: https://www.aiforhumans.show/   // SHOW LINKS //  Ideogram https://ideogram.ai/t/explore Ideogram 2.0 Launch Trailer https://x.com/i/status/1826277550798278804 Fluxstanza https://civitai.com/models/657252/fluxstanza?modelVersionId=735368 https://x.com/itspoidaman/status/1824957283803308536 FAL.ai https://fal.ai/models Political AI Image Controversy https://www.nytimes.com/2024/08/19/us/politics/trump-taylor-swift-ai-images.html Imagen-3 https://deepmind.google/technologies/imagen-3/ Procreate Boss Says They'll Never Use AI  https://www.theverge.com/2024/8/19/24223473/procreate-anti-generative-ai-pledge-digital-illustration-creatives OpenAI Fine Tuning Launched For GPT-4o (explain what this is…) https://openai.com/index/gpt-4o-fine-tuning/ Partnering with Conde Nast https://openai.com/index/conde-nast/ Matthew Berman SearchGPT Video https://youtu.be/DV9I_fu0ba8?si=Kcd4wb54wFgwaQuJ Unitree's 16k Humanoid Robot Goes Into Production  https://x.com/UnitreeRobotics/status/1789931753974517820 GoT AI Rave https://x.com/andr3_ai/status/1825600625754911091 Space Vets Children's Series:  https://www.storybookstudios.ai/space-vets Demis Hassabis on the Google Deep Mind Podcast https://x.com/GoogleDeepMind/status/1824447036847993292 Runway GEN-3 Turbo https://venturebeat.com/ai/runways-gen-3-alpha-turbo-is-here-and-can-make-ai-videos-faster-than-you-can-type/ Hedra 1.5 Character Generation https://x.com/hedra_labs/status/1824113944757457157  

The top AI news from the past week, every ThursdAI

Hey there, Alex here with an end of summer edition of our show, which did not disappoint. Today is the official anniversary of stable diffusion 1.4 can you believe it? It's the second week in the row that we have an exclusive LLM launch on the show (after Emozilla announced Hermes 3 on last week's show), and spoiler alert, we may have something cooking for next week as well!This edition of ThursdAI is brought to you by W&B Weave, our LLM observability toolkit, letting you evaluate LLMs for your own use-case easilyAlso this week, we've covered both ends of AI progress, doomerist CEO saying "Fck Gen AI" vs an 8yo coder and I continued to geek out on putting myself into memes (I promised I'll stop... at some point) so buckle up, let's take a look at another crazy week: TL;DR* Open Source LLMs * AI21 releases Jamba1.5 Large / Mini hybrid Mamba MoE (X, Blog, HF)* Microsoft Phi 3.5 - 3 new models including MoE (X, HF)* BFCL 2 - Berkley Function Calling Leaderboard V2 (X, Blog, Leaderboard)* NVIDIA - Mistral Nemo Minitron 8B - Distilled / Pruned from 12B (HF)* Cohere paper proves - code improves intelligence (X, Paper)* MOHAWK - transformer → Mamba distillation method (X, Paper, Blog)* AI Art & Diffusion & 3D* Ideogram launches v2 - new img diffusion king

Ckb Show : le podcast qui parle de Google
Nos favoris technologiques de l'été

Ckb Show : le podcast qui parle de Google

Play Episode Listen Later Aug 12, 2024 95:03


Bienvenue dans cet épisode estivale de votre podcast, dédié à nos coups de cœur de l'été ! Aujourd'hui, nous allons partager avec vous une sélection d'extensions, de sites web et d'outils boostés par l'intelligence artificielle qui nous ont marqué cet été. Que vous soyez à la recherche de nouvelles façons d'optimiser votre productivité, de découvrir des plateformes innovantes ou d'explorer des outils IA fascinants, cet épisode est fait pour vous. Vous allez découvrir ByPass Paywalls, Kozikaza MaxFocus et bien d'autres applications. Alors, installez-vous confortablement et préparez-vous à découvrir nos trouvailles estivales préférées, qui, nous l'espérons, vous inspireront autant qu'elles nous ont séduits ! 00:00:00 Introduction 00:07:15 Les extensions 00:07:15 Bypass Paywalls: https://github.com/iamadamdev/bypass-paywalls-chrome 00:15:17 MaxFocus: https://chromewebstore.google.com/detail/bnacincmbaknlbegecpioobkfgejlojp?hl=fr 00:18:45 Dictionnaire: https://chromewebstore.google.com/detail/nhbchcfeodkcblfpdjdhelcfbefefmag?hl=fr 00:21:40 Youtube Summary with ChatGPT & Claude: https://chromewebstore.google.com/detail/nmmicjeknamkfloonkhhcjmomieiodli 00:25:35 TextBlaze: https://blaze.today/ 00:31:36 ChromeSignBuilder: https://chromewebstore.google.com/detail/chrome-sign-builder/odjaaghiehpobimgdjjfofmablbaleem?utm_source=pocket_shared 00:34:14 Brisk Teaching: https://www.briskteaching.com/fr-fr 00:38:03 TranscripTonic: https://chromewebstore.google.com/detail/transcriptonic/ciepnfnceimjehngolkijpnbappkkiag 00:40:05 Les sites Web 00:40:35 kozikaza: https://www.kozikaza.com/kazaplan/new 00:43:09 spreadsimple: https://spreadsimple.com/ 00:46:05 Tally: https://tally.so/ 00:49:43 Bento3D.design: https://bento3d.design/ 00:51:50 https://www.oneusefulthing.org/ 00:54:36 Chromebook, le champion de la longévité 01:00:24 L'IA dans tout ça ? 01:02:09 Preceden: https://www.preceden.com 01:03:47 Ideogram.ai: https://ideogram.ai/ 01:06:34 Microsoft Designer: https://designer.microsoft.com/ 01:08:55 Seelab: https://seelab.ai/ 01:13:10 brev.ai: https://brev.ai/ 01:17:32 NotebookLM: https://www.notebooklm.google/ 01:22:10 Food Mood: https://www.mrcook.app/ 01:23:50 GenType: https://labs.google/gentype 01:25:30 Gamma.app: https://gamma.app/ 01:28:19 À tester : https://elevenlabs.io/ 01:31:15 À tester : https://www.rosebud.ai/ Nicolas Drolo : https://solo.to/mychromebook Alexandra Coutlée : https://www.lageekdeservice.com/

a16z
When AI Meets Art

a16z

Play Episode Listen Later Jul 30, 2024 43:20


On June 27th, the a16z team headed to New York City for the first-ever AI Artist Retreat at their office. This event brought together the builders behind some of the most popular AI creative tools, along with 16 artists, filmmakers, and designers who are exploring the capabilities of AI in their work.In this episode, we hear from the innovators pushing the boundaries of AI creativity. Joined by Anish Acharya, General Partner, and Justine Moore, Partner on the Consumer team, we feature insights from:Ammaar Reshi - Head of Design, ElevenLabsJustin Maier - Cofounder & CEO, CivitaiMaxfield Hulker - Cofounder & COO, CivitaiDiego Rodriguez - Cofounder & CTO, KreaVictor Perez - Cofounder & CEO, KreaMohammad Norouzi - Cofounder & CEO, IdeogramHang Chu - Cofounder & CEO, ViggleConor Durkan - Cofounder, UdioThese leaders highlight the surprising commonalities between founders and artists, and the interdisciplinary nature of their work. The episode covers the origin stories behind these innovative tools, their viral moments, and their future visions. You'll also hear about the exciting potential for AI in various creative modalities, including image, video, music, 3D, and speech.Keep an eye out for more in our series highlighting the founders building groundbreaking foundation models and AI applications for video, audio, photography, animation, and more.Learn more and see videos on artists leveraging AI at: a16z.com/aiart Find Ammaar on Twitter: https://x.com/ammaarLearn more about ElevenLabs: https://elevenlabs.ioFind Justin on Twitter: https://x.com/justmaierFind Max on LinkedIn: https://www.linkedin.com/in/maxfield-hulker-5222aa230/Learn more about Civitai: https://civitai.comFind Diego on Twitter: https://x.com/asciidiego?lang=enFind Victor on Twitter: https://x.com/viccpoesLearn more about Krea: https://www.krea.ai/homeFind Mohammed on Twitter: https://x.com/mo_norouziLearn more about Ideogram: https://ideogram.ai/t/exploreFind Conor on Twitter: https://x.com/conormdurkanLearn more about Udio: https://www.udio.com/homeFind Hang on Twitter: https://x.com/chuhang1122Learn more about Viggle: https://viggle.ai/ Stay Updated: Let us know what you think: https://ratethispodcast.com/a16zFind a16z on Twitter: https://twitter.com/a16zFind a16z on LinkedIn: https://www.linkedin.com/company/a16zSubscribe on your favorite podcast app: https://a16z.simplecast.com/Follow our host: https://twitter.com/stephsmithioPlease note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.  

AI + a16z
The Future of Image Models Is Multimodal

AI + a16z

Play Episode Listen Later Jun 7, 2024 37:17


In this episode, Ideogram CEO Mohammad Norouzi joins a16z General Partner Jennifer Li, as well as Derrick Harris, to share his story of growing up in Iran, helping build influential text-to-image models at Google, and ultimately cofounding and running Ideogram. He also breaks down the differences between transformer models and diffusion models, as well as the transition from researcher to startup CEO.Here's an excerpt where Mohammad discusses the reaction to the original transformer architecture paper, "Attention Is All You Need," within Google's AI team:"I think [lead author Asish Vaswani] knew right after the paper was submitted that this is a very important piece of the technology. And he was telling me in the hallway how it works and how much improvement it gives to translation. Translation was a testbed for the transformer paper at the time, and it helped in two ways. One is the speed of training and the other is the quality of translation. "To be fair, I don't think anybody had a very crystal clear idea of how big this would become. And I guess the interesting thing is, now, it's the founding architecture for computer vision, too, not only for language. And then we also went far beyond language translation as a task, and we are talking about general-purpose assistants and the idea of building general-purpose intelligent machines. And it's really humbling to see how big of a role the transformer is playing into this."Learn more:Investing in IdeogramImagenDenoising Diffusion Probabilistic ModelsFollow everyone on X:Mohammad NorouziJennifer LiDerrick Harris Check out everything a16z is doing with artificial intelligence here, including articles, projects, and more podcasts.

da Brand a Friend
#323 - Gli Ostacoli Più Grandi

da Brand a Friend

Play Episode Listen Later Jun 2, 2024 11:45


Se mi dovessero intervistare e chiedere quali vedo essere gli ostacoli più grandi per i piccoli imprenditori indipendenti, risponderei con: 1) consistenza e continuità2) relazioni - promozione - outreach3) mettersi in discussioneSe mi domandi cosa ho imparato nel mio percorso e come ho risolto questi tre ostacoli quando li ho incontrati, ti direi:1) Consistenza e continuitàQuesto è l'ostacolo base.Mi sono seduto e ho pensato che non sarebbe successo nullaInfatti nulla, di rumoroso o visibile è successo, ma il mio business ha smesso di crescere ogni volta che non ho dato continuità ad un progetto.Questo è l'ostacolo più frequente e comune. Nel mio piccolo, lo ho risolto liberando sempre più tempo nella mia vita - facendo in modo ad essere io a dettare i tempi e non le circostanze della vita.2) Relazioni - promozione - outreach Questo è l'ostacolo più difficile da superare, e allo stesso tempo il più semplice.È il coltivare le relazioni.L'interessarsi agli altri.Il farsi sentire.Lo stringere rapporti e relazioni con altre persone interessanti nel tuo spazio.Commentare, ingaggiare, domandare. È anche quello che da i frutti più belli e inaspettati, a volte in tempi assai breviL'ho risolto uscendo dalla mia zona di comfort e facendo cose che mi davano la possibilità di scambiare in maniera diretta con persone. 3) Mettersi in discussioneQuesto è l'ostacolo invisibile. Finché è troppo tardi per poterlo affrontare.Poca disponibilità a rivedere i propri passi e le proprie scelte. Anche se il fare errori e di conseguenza cambiare / correggere la direzione è la strada  migliore per crescere e migliorare.Continua a fare il compitino perché così qualcuno ci ha detto, pensavamo che…, abbiamo imparato così.Continuare a innaffiare la stessa pianta, anche quando fa fiori e non da frutti, è una perdita di tempo e di denaro.Di tanto in tanto è utile farsi un piccolo bagno di umiltà e chiedere (o addirittura pagare) che qualcuno esperto ed esterno a noi analizzi criticamente ciò che facciamo e suggerisca strategie e tattiche per migliorare le nostre prestazioni.Sono di natura uno a cui piace mettersi in discussione. Mi piace sperimentare e provare nuove strade e soluzioni.Può succedere però che uno si impigrisca. …che incroci le dita e speri che le cose vadano migliorando… ma no, non è saggio fare così.Bisogna costantemente mettersi in discussione e quando le cose non vanno e non migliorano, farsi domande e fare cambiamenti… non aspettare e vedere._______________Info Utili• Sostieni questo podcast:Ottieni feedback, ricevi consigli sul tuo progetto onlinehttps://Patreon.com/Robin_Good•  Musica di questa puntata:"Let's Go Surfing" by Joakim Karud disponibile su Bandcamp•  Nella foto di copertina:Continuità, Relazioni, Mettersi in Discussione. Immagine generata da Ideogram.ai• Dammi feedback:Critiche, commenti, suggerimenti, idee e domande unendoti al gruppo Telegram https://t.me/@RobinGoodPodcastFeedback• Ascolta e condividi questo podcast:https://www.spreaker.com/show/dabrandafriend• NUOVO! Archivio completo organizzato per temi:https://start.me/p/kxENzk/da-brand-a-friend-archivio-podcast• Seguimi su Telegram:https://t.me/RobinGoodItalia.

da Brand a Friend
#319 - IA: Dove Batterla

da Brand a Friend

Play Episode Listen Later May 5, 2024 14:51


Un paio di settimane fa ho prodotto una vera e propria canzone, usando una nuova app di intelligenza artificiale che si chiama Udio. Ecco il pezzo. Si chiama “La Puerta Azul” e l'ho creato per l'omonimo beach club qui sull'isola di Holbox, dove suono come DJ. Non potevo credere alle mie orecchie. Troppo bello per essere vero. Considerando il fatto che io non sono musicista, non ho studiato musica, e non ho investito più di 5-6 ore per realizzare questo pezzo direi che c'è quasi da preoccuparsi.E così ho fatto.Mi sono domandato: ma se adesso, così come sta succedendo con le immagini, produrre canzoni diventa teoricamente alla portata di chiunque, che succederà? Come farà chi fa il fotografo, il pittore, il musicista o l'astrologo a non trovarsi improvvisamente rimpiazzato da una IA (intelligenza artificiale)?Quando c'è sovrabbondanza di qualcosa, come si fa a rimanere rilevanti?La risposta sta:a) nel saper riconoscere e distinguere - in un mare di opzioni - quelle veramente speciali. Questo da solo è un mestiere per il futuro (in qualsiasi campo). b) nel rifinire e innovare la propria “arte” in modo che esprima emozioni, capacità e idee che l'IA non potrà imitare, perché sono esclusivamente tue. _______________Info Utili• Sostieni questo podcast:Ottieni feedback, ricevi consigli sul tuo progetto online•  Entra nella comunità di imprenditori indipendenti di Robin Goodhttps://Patreon.com/Robin_Good•  Musica di questa puntata:"Bittersweet" by baskaaT - dsponibile su Soundcloud•  Nella foto di copertina:Immagine generata dall'IA. Ideogram.ai. Mag 2024• Dammi feedback:Critiche, commenti, suggerimenti, idee e domande unendoti al gruppo Telegram https://t.me/@RobinGoodPodcastFeedback• Ascolta e condividi questo podcast:https://www.spreaker.com/show/dabrandafriend• NUOVO! Archivio completo organizzato per temi:https://start.me/p/kxENzk/da-brand-a-friend-archivio-podcast• Seguimi su Telegram:https://t.me/RobinGoodItalia.

SNAP - Architettura Imperfetta
Intelligenza democratizzata - Ideogram OICE nVidia | 257

SNAP - Architettura Imperfetta

Play Episode Listen Later Apr 19, 2024 29:24


Bentornati su Snap!Iniziamo con una prova sul campo di Ideogram, un nuovo tool di immagini generative ed una bella carrellata di app che collaborano con l'AI per la progettazioni d'interni, consigliata dal Pro Snapper Fabri LT.La parte centrale è da ascoltare da seduti perchè riguarda il BIM ed i bandi dei lavori pubblici inquadrati dal rapporto OICE per l'anno 2023.Passiamo oltre e guardiamo questo 2024, con Apple che stringe accordi per addestrare la sua AI e nVidia che crea un supercomputer a disposizione degli studenti.Buon ascolto!—>

Everyday VOpreneur
Boost Your Marketing With AI: Elevate Your Graphics And Images with Mandi Kaye

Everyday VOpreneur

Play Episode Listen Later Apr 11, 2024 41:49


Mandi Kaye joins Marc Scott to discuss AI Image Generation and how voice actors can use it in their marketing efforts. Mandi shares her experience and insights on using various AI tools for image generation, including Dall-E, Ideogram and Leonardo. They discuss the evolution of AI technology, the benefits and limitations of different tools, and the importance of prompts in generating desired images. Mandi shares how she uses AI image generation to create social media posts and marketing materials for her voiceover business. She explains that she starts by writing the text she wants to accompany the image and then asks ChatGPT to generate an image that matches. She iterates on the generated images until she finds the perfect one. Mandi also discusses the importance of creating detailed prompts to get the desired results. She shares examples of prompts she has used and emphasizes the need to be specific and thorough. Potential concerns about using AI-generated content is also discussed. CONNECT WITH MANDI KAYE Mandi Kaye Website - https://mandikaye.com Mandi Kaye on Instagram - https://www.instagram.com/themandikaye Mandi Kaye on LinkedIn - https://www.linkedin.com/in/mandisorensen   Marc Scott on Instagram - @marcscott AI RESOURCES MENTIONED IN THIS EPISODE Dall-E - https://chat.openai.com Canva - https://canva.com Midjourney - https://www.midjourney.com Adobe Firefly - https://firefly.adobe.com Ideogram - https://ideogram.ai Leonardo - https://leonardo.ai RESOURCES FOR VOICE ACTORS * The VOpreneur Guide to Testimonials Visit https://vopreneur.com/testimonials * Get an instant $25 credit when you sign up for VoiceZam Visit https://voicezam.com/marcscott * For voice over services: Visit https://marcscottvoiceover.com * Want VOpreneur Swag? Visit https://teespring.com/stores/vopreneur * Join the VOpreneur Facebook Group Visit https://facebook.com/groups/vopreneur EVERYDAY VOPRENEURS IN THIS EPISODE * Thanks to "Uncle Roy" for production assistance! Visit https://antlandproductions.com * Thanks to Christy Harst for VO contributions! Visit https://christyharst.com * Thanks to Krysta Wallrauch for VO contributions! Visit https://krystawallrauch.com If you need guidance with your voice over business or learning how to more effectively market, I can help. Book a 15 minute free consultation with me to discuss your specific needs. Book Your Consult KEY TAKEAWAYS AI image generation offers exciting possibilities for voice actors. Different AI tools have different strengths and weaknesses, and it's important to choose the right tool for the desired outcome. Prompts play a crucial role in generating desired images, and experimenting with prompts can lead to better results. AI image generation can be used for various applications, such as creating quote graphics and social media graphics for voiceover businesses. AI image generation can be used to create social media posts and marketing materials for businesses Writing detailed prompts is crucial to getting the desired results from AI image generation AI-generated content should be used with real content to avoid any potential issues AI image generation can save time and help with brainstorming and ideation  

Marketing Against The Grain
Matt Wolfe Ranks The Best AI Tools For Marketers In 2024

Marketing Against The Grain

Play Episode Listen Later Apr 9, 2024 38:22


Ep. 215 What are the hottest and most important Ai apps to use in your marketing today? Kipp and guest Matt Wolfe (entrepreneur and YouTuber) dive into the intricate world of AI marketing tools, exploring what it means to effectively integrate AI into your toolkit. Learn more about how Hume can decipher the subtleties of human emotion, the creative potential of using Suno for song creation, and the practical uses of both Recraft  and Ideogram for crafting compelling marketing visuals. Check out Matt's new podcast The Next Wave here: https://link.chtbl.com/uqnVaUip Mentions Hume https://www.hume.ai/ Suno https://www.suno.ai/ Recraft https://www.recraft.ai/ Ideogram https://ideogram.ai/ We're on Social Media! Follow us for everyday marketing wisdom straight to your feed YouTube: ​​https://www.youtube.com/channel/UCGtXqPiNV8YC0GMUzY-EUFg  Twitter: https://twitter.com/matgpod  TikTok: https://www.tiktok.com/@matgpod  Join our community https://landing.connect.com/matg Thank you for tuning into Marketing Against The Grain! Don't forget to hit subscribe and follow us on Apple Podcasts (so you never miss an episode)! https://podcasts.apple.com/us/podcast/marketing-against-the-grain/id1616700934   If you love this show, please leave us a 5-Star Review https://link.chtbl.com/h9_sjBKH and share your favorite episodes with friends. We really appreciate your support. Host Links: Kipp Bodnar, https://twitter.com/kippbodnar   Kieran Flanagan, https://twitter.com/searchbrat  ‘Marketing Against The Grain' is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Produced by Darren Clarke.

The top AI news from the past week, every ThursdAI

Hello hello everyone, happy spring! Can you believe it? It's already spring! We have tons of AI news for you to cover, starting with the most impactful one, did you already use Claude 3? Anthropic decided to celebrate Claude 1's birthday early (which btw is also ThursdAI's birthday and GPT4 release date, March 14th, 2023) and gave us 3 new Clauds! Opus, Sonnet and Haiku. TL;DR of all topics covered: * Big CO LLMs + APIs*

AI For Humans
OpenAI vs Elon, Anthropic's Claude 3 Is Great & AIs Debate Toilet Paper | Ep47

AI For Humans

Play Episode Listen Later Mar 7, 2024 93:23


This week… Elon Musk sues OpenAI, Anthropic's Claude 3 is the best LLM we've used, Google's Sergey Brin knows they screwed up & AI tax bots gone very wrong. Plus, Gavin dives into text-to-image newness with Ideogram 1.0, Kevin shows off Dust3r which does simple 3D modeling, another dancing robot and a group of hackers takes on Humane's AI pin.  AND THEN…it's the return of the AI For Humans AI Debate! We pit OpenAI's GPT-4 against the newly released Claude 3 and the results will both surprise and, dare we say, SHOCK YOU TO YOUR CORE. This week's AI co-host is one of our two debaters, Dr. Cornelius “Corny” Quckenbush, who has come to take on GPT-4 and also tell us of his deep love of rubber ducks. It's an endless cavalcade of ridiculous and informative AI news, AI tools, and AI entertainment cooked up just for you. Follow us for more AI discussions, AI news updates, and AI tool reviews on X @AIForHumansShow Join our vibrant community on TikTok @aiforhumansshow For more info, visit our website at https://www.aiforhumans.show/   /// Show links /// Anthropic's Claude 3 https://claude.ai/ Elon Vs OpenAI https://www.nytimes.com/2024/03/02/technology/elon-musk-openai-lawsuit-microsoft-research.html Cecilia Ziniti's Twitter Thread https://twitter.com/CeciliaZin/status/1763849318396752151 Claude 3 Is Here https://x.com/AnthropicAI/status/1764653830468428150?s=20 Claude3 Has Awareness of Doing a Test https://x.com/alexalbert__/status/1764722513014329620?s=20 Sergey Brin: We Messed Up https://fortune.com/2024/03/04/sergey-brin-google-definitely-messed-up-gemini-image-generation/ Laurie Anderson Brings Lou Reed Back https://www.theguardian.com/music/2024/feb/28/laurie-anderson-ai-chatbot-lou-reed-ill-be-your-mirror-exhibition-adelaide-festival?utm_source=aisecret.us&utm_medium=Aisecret.us&utm_campaign=Daily AI Political Generated Images That Aren't Real https://www.theguardian.com/us-news/2024/mar/04/trump-ai-generated-images-black-voters Washington Post: Tax Chatbots Are Screwing Up https://www.washingtonpost.com/technology/2024/03/04/ai-taxes-turbotax-hrblock-chatbot/ AI App Helps Detect Ear Infections In Children https://www.cbsnews.com/pittsburgh/news/ai-smartphone-app-diagnose-ear-infections-pittsburgh/ Whomane: Open Source AI Pin https://x.com/kodjima33/status/1764472814353183199?s=20 Expressive Whole Body Control for Humanoid Robots https://youtu.be/UGA9YAg3e-M?si=cYqM30UoOPIsNuzU WTFaldo (Waldo Animation) https://x.com/CitizenPlain/status/1764763312107970592?s=20 Dust3r https://dust3r.europe.naverlabs.com/ Ideogram https://ideogram.ai/t/explore  

Unsupervised Learning
Ep 29: Salesforce AI CEO Clara Shih on Future of Slack, How Gucci Uses AI and Working with Marc Benioff

Unsupervised Learning

Play Episode Listen Later Mar 6, 2024 52:37


There's an ongoing debate about where the most value will accrue in AI between incumbents and startups. Of the incumbents, few have shipped product faster than SalesforceAI. Today on Unsupervised Learning we had on Clara Shih, CEO of SalesforceAI and one of Time Magazine's 100 Most Influential People in AI.  (0:00) intro(0:50) work practices that will become irrelevant(1:37) revolutionizing reply recommendations and case summaries(4:57) newest Salesforce products(5:53) structuring teams(7:22) engineering trust into AI products(11:58) combining in-house models with ChatGBT(13:33) Gucci's AI adoption(16:01) how does Salesforce choose who to share their data with?(20:29) AI costs(26:29 creating unique voices for brands(27:45) AI incumbents vs. startups(29:54) what Clara would build if she had the time(32:28) the future of Slack(35:55) what percent of customer support questions can be answered by AI?(38:37) over-hyped/under-hyped(39:32) working with Mark Benioff(40:46) Jacob and Pat debrief(44:42) Slack is the perfect interface for generative AI(46:10) Abridge investment(48:15) Ideogram investment With your co-hosts:  @jacobeffron  - Partner at Redpoint, Former PM Flatiron Health  @patrickachase  - Partner at Redpoint, Former ML Engineer LinkedIn  @ericabrescia  - Former COO Github, Founder Bitnami (acq'd by VMWare)  @jordan_segall  - Partner at Redpoint

This Week in XR Podcast
This Week In XR March 1st, 2024 ft. Hugh Forrest, Co-President & Chief Programming Officer at SXSW

This Week in XR Podcast

Play Episode Listen Later Mar 1, 2024 42:03


Charlie and Ted are without Rony, who is driving across the country. We'll convene at SXSW in Austin to do our show live on stage with Joanna Popper in a session titled "XR in the Age of Vision Pro" on March 12th. In the news, $675M for Figure AI's humanoid robots, powered by ChatGPT, and a new text-to-video tool, Ideogram. Our guests this week are Hugh Forrest, head of programming for SXSW and Blake Kammerdiner, head of XR at SXSW who give us a wide-ranging look at this year's show. To say Ted and Charlie are fanboys would be putting it mildly. Thank you to our sponsor, Zappar!Don't forget to like, share, and follow for more! Follow us on all socials @ThisWeekInXR!https://linktr.ee/thisweekinxr Hosted on Acast. See acast.com/privacy for more information.

The top AI news from the past week, every ThursdAI

Happy leap year day everyone, very excited to bring you a special once-in-a-4 year edition of ThursdAI

ceo new york tiktok friends europe english google ai apple moving training french new york times deep chinese european german spanish microsoft italian open chatgpt hong kong mcdonald tree curse cheers reddit mac stanford cat billion honestly dune adams vibes mark zuckerberg ego mighty berkeley communicate cto programming leap messy folks similar swift mj react excel transformers largest includes wordpress genie function openai user gemini gem fusion nvidia complaints rust references tumblr wing api trained ye documents stack open source ak mojo python trillion turbo gpt gorilla playground aws ml lama alibaba 2d github mayo clinic llama analyze dua lipa iso ds apis foundational transformer hermes sum javascript azure existential apache tl emo imagen sora copilot daly prompt llm cpu yum gpu beautifully 3b hug orca dali modular vector midjourney phi coherence avatars instruct diffusion rag leap year texture guerrilla 7b 1k automatically pca lemme fai perplexity tldr aditya mtb grok ess fine tuning lms satya fatih retrieval lm yam yee json jaw sundar pichai sota ouroboros representations tropic mistral stable diffusion typescript chunks jammer clippy automattic olam a16z year special seamlessly tensorflow abacus le chat nissen pratik axolotl dpo junaid prateek pytorch cohere 15b open source ai matryoshka tpu mixpanel dicta larynx llvm groq loras dimensionality chris lattner neurips mira murati sft huggingface gemini pro jema rlhf entropic cerebras mrl ideogram code llama hrithik adithya scipy technium gemini ultra weaviate olami matplotlib mlir suhail doshi rohf partik john durbin
La mia vita spaziale
Ideogram: immagini AI gratis

La mia vita spaziale

Play Episode Listen Later Feb 24, 2024 4:32


In questo episodio di "La mia vita spaziale", mi addentro nel mondo affascinante dell'intelligenza artificiale per svelarvi Ideogram, un'applicazione gratuita che consente a chiunque di trasformare idee in immagini, sfruttando il potere quasi magico dell'AI.

Blogs y Blogging: El PODCAST de Blogpocket
HECHO CON BLOQUES #48: Especial herramientas de IA generativa

Blogs y Blogging: El PODCAST de Blogpocket

Play Episode Listen Later Feb 23, 2024 9:45


Este episodio de HECHO CON BLOQUES es un especial sobre herramientas de IA generativa. Se muestran varias herramientas como Opus Clip, Eleven Labs, Ideogram y Suno. También se menciona el uso responsable de la IA y se hace referencia a un manifiesto sobre la implementación responsable de la IA que hemos suscrito en Blogpocket. [Encontrarás la transcripción completa y los enlaces a los recursos citados, en Blogpocket.com]

How to Sell Your Stuff on Etsy
Ep 115 | Turn Your Handmade Business into a Multi-Stream Machine – with The Product Boss

How to Sell Your Stuff on Etsy

Play Episode Listen Later Feb 8, 2024 53:13


Have you ever heard the phrase— “Don't put all your eggs in the Etsy basket?” Today the Product Boss herself is joining us for an inspiring conversation about how you can scale your product based business into multiple streams of income. Listen in to learn what you can do to safeguard and grow your Etsy business into a multi-stream machine from the best in the business. **“How to Sell Your Stuff on Etsy” is not affiliated with or endorsed by Etsy.com STUFF I MENTIONED: Check out the “How to Sell Your Stuff” Youtube Channel: THANKS for your support!!!   FREE Best Seller Secret 5 Day Challenge (Feb 12-16, 2024): https://www.theproductboss.com/etsy Multi Stream Machine (available 2/14/24): https://theproductboss.samcart.com/referral/msm/QUGSlkiR2rCqAnii The Product Boss Podcast: https://www.theproductboss.com/podcast Instagram: http://instagram.com/theproductboss Facebook: https://www.facebook.com/theproductboss WHAT'S HAPPENING NOW: ⭐The A.I. Print on Demand LIVE Workshop just happened on 1/17/24! Get a copy of the recording, prompts for your own mockups + POD designs, tutorials for Midjourney, Ideogram, and Dalle-3 and MORE (Use code POD50 to save $50): https://www.howtosellyourstuff.com/ai-POD-workshop-enrollment ⭐ Get my free list of 100 in Demand Micro-Niche Keywords: https://www.howtosellyourstuff.com/Micro-Niche-Demand  ⭐ Join the $4.99 monthly subscription to the In Demand Micro-Niche Keywords List: https://www.howtosellyourstuff.com/micro-niche-member  (You'll get ongoing access to the ever growing of list of hundreds--soon to be thousands-- of micro-niche opportunities on Etsy!)  ---------------------------------------------   ⭐Book a one-on-one Etsy coaching session with Lizzie: https://www.howtosellyourstuff.com/coaching ⭐Apply to be a Podcast Guest: https://bit.ly/48hFD8X Find me on Instagram and TikTok @HowtoSellYourStuff   FREE ETSY MASTERCLASS: https://www.howtosellyourstuff.com/masterclass FREE PDF DOWNLOAD: “4 Strategies I Used to Grow My Etsy Shop from $25 to $6000k/month”: https://www.howtosellyourstuff.com/site/4-strategies-opt-in   Grab my UPDATED Etsy Course for physical product sellers: “Listings that Sell 2.0” and learn how to skyrocket your Etsy business: https://www.howtosellyourstuff.com/etsy-listings-that-sell    ----- HOW TO SELL YOUR STUFF WEBSITE: https://www.howtosellyourstuff.com/ HOW TO SELL YOUR STUFF INSTAGRAM: https://www.instagram.com/howtosellyourstuff/ HOW TO SELL YOUR STUFF SHOWNOTES: https://www.howtosellyourstuff.com/blog/115 ------- THIS EPISODE IS SPONSORED BY: My most popular freebie— the FREE PDF download where I share the “4 Strategies I Used to Grow My Etsy Shop from $25/month to $6000+/month.” Grab a copy and start leveling up your Etsy shop: https://www.howtosellyourstuff.com/site/4-strategies-opt-in AND   Do you use special fonts, graphics, svgs, or other digital goods to create your products or run your Etsy business? You NEED Creative Fabrica! Creative Fabrica is a website where you can access UNLIMITED digital goods for just $9 per month. They have over 6 MILLION fonts, graphics, and other digital resources that you will gain full access to. (It's essentially the top Etsy seller's best kept secret!) AND on top of all that Creative Fabrica discovered this podcast and reached out to me because they wanted to offer you guys a special little perk: you can now get a one-month free trial for up to 10 product downloads to test drive it and see if it's a good fit for you. Learn more at: https://www.creativefabrica.com/ref/2877703 (Now through 2/14/24--- get the FULL YEAR for only $47! It's a crazy deal!)   *Some of the links above are affiliate links which means I'll receive a commission if you purchase through my link, at no extra cost to you. You can see my affiliate disclosure here: https://www.howtosellyourstuff.com/affiliate-disclosure

UiPath Daily
Ideogram's Design Revolution: AI Image Generator Introduces (Good) Custom Typography!

UiPath Daily

Play Episode Listen Later Feb 2, 2024 5:41


Experience a design revolution with Ideogram's AI Image Generator, showcasing impressive custom typography to elevate your visual content. Delve into the creative possibilities and discover a new era in image generation. Get on the AI Box Waitlist: AIBox.ai Join our ChatGPT Community: Facebook Group Follow me on Twitter: Jaeden's Twitter

ChatGPT: OpenAI, Sam Altman, AI, Joe Rogan, Artificial Intelligence, Practical AI
Custom Typography Brilliance: Ideogram Launches AI Image Generator for Stunning Visuals!

ChatGPT: OpenAI, Sam Altman, AI, Joe Rogan, Artificial Intelligence, Practical AI

Play Episode Listen Later Feb 2, 2024 5:41


Immerse yourself in the brilliance of custom typography with Ideogram's AI Image Generator, setting a new standard for stunning visual content. Explore the possibilities and unleash your creativity in the world of design. Get on the AI Box Waitlist: AIBox.ai Join our ChatGPT Community: Facebook Group Follow me on Twitter: Jaeden's Twitter

AI for Non-Profits
Ideogram's Visionary Tool: AI Image Generator Debuts with (Good) Custom Typography!

AI for Non-Profits

Play Episode Listen Later Feb 2, 2024 5:41


Step into the future of design with Ideogram's AI Image Generator, introducing visionary custom typography for impactful visuals. Delve into the innovative features and redefine your approach to creating stunning images. Get on the AI Box Waitlist: AIBox.ai Join our ChatGPT Community: Facebook Group Follow me on Twitter: Jaeden's Twitter

The Elon Musk Podcast
Crafting Visual Stories: Ideogram's AI Image Generator with Custom Typography

The Elon Musk Podcast

Play Episode Listen Later Feb 2, 2024 7:13


Join me in unpacking Ideogram's groundbreaking AI image generator, a tool that redefines visual storytelling with its superior custom typography features, set to transform content creation. Invest in AI Box: https://Republic.com/ai-box Get on the AI Box Waitlist: https://AIBox.ai/ AI Facebook Community

Open AI
Elevate Your Designs: Ideogram Launches AI Image Generator with Custom Typography!

Open AI

Play Episode Listen Later Feb 2, 2024 5:41


Join the design revolution with Ideogram's latest innovation – an AI Image Generator that boasts impressive custom typography. Uncover the potential to transform your visuals and create stunning content effortlessly. Get on the AI Box Waitlist: AIBox.ai Join our ChatGPT Community: Facebook Group Follow me on Twitter: Jaeden's Twitter

How to Sell Your Stuff on Etsy
Ep 114 | Sewing Hobby Turned College Etsy Side Hustle

How to Sell Your Stuff on Etsy

Play Episode Listen Later Feb 1, 2024 53:03


Miss Maddy learned how to sew as a child and fell in love with the hobby. As a teenager she turned her hobby into an Etsy side-hustle that helped her pay the bills all through college. Today her Cat Costume Shop has charmed almost 6,000 customers and continued to bring her joy, smiles, and a wonderful supplemental income. You're going to absolutely adore Maddy and her story! **“How to Sell Your Stuff on Etsy” is not affiliated with or endorsed by Etsy.com   STUFF I MENTIONED: Where to find Maddy: Etsy: https://www.etsy.com/shop/MissMaddyMakes Instagram: @miss.maddy.makes   WHAT'S HAPPENING NOW: ⭐The A.I. Print on Demand LIVE Workshop just happened on 1/17/24! Get a copy of the recording, prompts for your own mockups + POD designs, tutorials for Midjourney, Ideogram, and Dalle-3 and MORE (Use code POD50 to save $50): https://www.howtosellyourstuff.com/ai-POD-workshop-enrollment ⭐ Get my free list of 100 in Demand Micro-Niche Keywords: https://www.howtosellyourstuff.com/Micro-Niche-Demand  ⭐ Join the $4.99 monthly subscription to the In Demand Micro-Niche Keywords List: https://www.howtosellyourstuff.com/micro-niche-member  (You'll get ongoing access to the ever growing of list of hundreds--soon to be thousands-- of micro-niche opportunities on Etsy!)  ---------------------------------------------   ⭐Book a one-on-one Etsy coaching session with Lizzie: https://www.howtosellyourstuff.com/coaching ⭐Apply to be a Podcast Guest: https://bit.ly/48hFD8X Find me on Instagram and TikTok @HowtoSellYourStuff   FREE ETSY MASTERCLASS: https://www.howtosellyourstuff.com/masterclass FREE PDF DOWNLOAD: “4 Strategies I Used to Grow My Etsy Shop from $25 to $6000k/month”: https://www.howtosellyourstuff.com/site/4-strategies-opt-in   Grab my UPDATED Etsy Course for physical product sellers: “Listings that Sell 2.0” and learn how to skyrocket your Etsy business: https://www.howtosellyourstuff.com/etsy-listings-that-sell    ----- HOW TO SELL YOUR STUFF WEBSITE: https://www.howtosellyourstuff.com/ HOW TO SELL YOUR STUFF INSTAGRAM: https://www.instagram.com/howtosellyourstuff/ HOW TO SELL YOUR STUFF SHOWNOTES: https://www.howtosellyourstuff.com/blog/114 ------- THIS EPISODE IS SPONSORED BY: My business and Etsy coaching services:  Sometimes we just need a pair of expert eyes to help us see our path forward more clearly! Whether you need help troubleshooting in your Etsy shop, pivoting to a new product, expanding your business to sell courses, and so much more-- you can hire me by the hour to provide a recorded zoom coaching session. We will work together to figure out your next steps so you can work smarter, not harder!  Book your session today: https://www.howtosellyourstuff.com/coaching AND My Customer Service Templates & Mini-Course: Learn my exact customer service strategy AND get access to over 20 templates of my word-for-word responses to customers in everyday and difficult situations:  https://www.howtosellyourstuff.com/offers/wUXKPzRG/checkout

The Elon Musk Podcast
Crafting Visual Stories: Ideogram's AI Image Generator with Custom Typography

The Elon Musk Podcast

Play Episode Listen Later Feb 1, 2024 6:18


Join me in unpacking Ideogram's groundbreaking AI image generator, a tool that redefines visual storytelling with its superior custom typography features, set to transform content creation. Invest in AI Box: https://Republic.com/ai-box Get on the AI Box Waitlist: https://AIBox.ai/ AI Facebook Community

Midjourney
Typographic Brilliance: Ideogram's AI Image Generator Unveiled with Custom Finesse

Midjourney

Play Episode Listen Later Jan 28, 2024 6:18


In this episode, we witness typographic brilliance as Ideogram unveils its AI image generator, showcasing custom finesse. Join me for a solo discussion, where we explore the brilliance of custom typography and the innovative aspects introduced by this tool in the world of image generation. Invest in AI Box: ⁠https://Republic.com/ai-box⁠ Get on the AI Box Waitlist: ⁠https://AIBox.ai/⁠ ⁠AI Facebook Community Learn About ChatGPT Learn About AI at Tesla

How to Sell Your Stuff on Etsy
Ep 113 | Niching Down Strikes Again--Over 750 sales in 2 Years-- with Linda Sortino

How to Sell Your Stuff on Etsy

Play Episode Listen Later Jan 25, 2024 54:10


Does SEO matter? Yes. Do great pictures matter? Yes. But even MORE than all the typical Etsy advice--- finding DEMAND and niching down to serve that demand will help new (and old) sellers win every time! This week I'm interviewing Linda Sortino who has made over 700 sales selling bead boards. You may have never heard of this product---  but her customers can't wait to get their hands on one. Listen in to hear how serving a very specific customer and niche can build an incredible Etsy business. **“How to Sell Your Stuff on Etsy” is not affiliated with or endorsed by Etsy.com STUFF I MENTIONED: Linda's Favorite Episodes: #70 Print on Demand Insight, Inspo, and Tips You Won't Want to Miss: https://www.howtosellyourstuff.com/blog/best-pod-for-etsy #102 Fast Success on Etsy in a “Saturated” Niche: https://www.howtosellyourstuff.com/blog/102   Where to find Linda: Blog: www.comebeadwithme.com   Product Website: https://beadboardenvy.com/ https://www.facebook.com/comebeadwithlinda/ https://www.pinterest.com/beadwithlinda/   WHAT'S HAPPENING NOW: ⭐The A.I. Print on Demand LIVE Workshop just happened on 1/17/24! Get a copy of the recording, prompts for your own mockups + POD designs, tutorials for Midjourney, Ideogram, and Dalle-3 and MORE (Use code POD50 to save $50): https://www.howtosellyourstuff.com/ai-POD-workshop-enrollment ⭐ Get my free list of 100 in Demand Micro-Niche Keywords: https://www.howtosellyourstuff.com/Micro-Niche-Demand  ⭐ Join the $4.99 monthly subscription to the In Demand Micro-Niche Keywords List: https://www.howtosellyourstuff.com/micro-niche-member  (You'll get ongoing access to the ever growing of list of hundreds--soon to be thousands-- of micro-niche opportunities on Etsy!)  ---------------------------------------------   ⭐Book a one-on-one Etsy coaching session with Lizzie: https://www.howtosellyourstuff.com/coaching ⭐Apply to be a Podcast Guest: https://bit.ly/48hFD8X Find me on Instagram and TikTok @HowtoSellYourStuff   FREE ETSY MASTERCLASS: https://www.howtosellyourstuff.com/masterclass FREE PDF DOWNLOAD: “4 Strategies I Used to Grow My Etsy Shop from $25 to $6000k/month”: https://www.howtosellyourstuff.com/site/4-strategies-opt-in   Grab my UPDATED Etsy Course for physical product sellers: “Listings that Sell 2.0” and learn how to skyrocket your Etsy business: https://www.howtosellyourstuff.com/etsy-listings-that-sell    ----- HOW TO SELL YOUR STUFF WEBSITE: https://www.howtosellyourstuff.com/ HOW TO SELL YOUR STUFF INSTAGRAM: https://www.instagram.com/howtosellyourstuff/ HOW TO SELL YOUR STUFF SHOWNOTES: https://www.howtosellyourstuff.com/blog/113 ------- THIS EPISODE IS SPONSORED BY: 100 KEYWORDS: Get the FREE list of 100 Keywords in various micro niches that all have demand without crazy competition ➡️ https://www.howtosellyourstuff.com/Micro-Niche-Demand  AND Listings that Sell 2.0 Learn all the secrets to build a 6 figure physical product shop with my flagship course Listings that Sell 2.0: https://www.howtosellyourstuff.com/etsy-listings-that-sell

AI Breakdown
Creative Control: Ideogram's AI Image Generator and Custom Typography

AI Breakdown

Play Episode Listen Later Jan 21, 2024 6:18


In this episode, we take a closer look at Ideogram's AI image generator, highlighting the exciting feature of customizable typography that grants users unprecedented creative control over their digital visuals. Invest in AI Box: ⁠https://Republic.com/ai-box⁠ Get on the AI Box Waitlist: ⁠https://AIBox.ai/⁠ ⁠AI Facebook Community Learn About ChatGPT Learn About AI at Tesla

How to Sell Your Stuff on Etsy
Ep 112 | Bailey has earned over $1 million in 3 years selling PNGs on Etsy – with Bailey Designed Co

How to Sell Your Stuff on Etsy

Play Episode Listen Later Jan 18, 2024 48:07


This digital product success story will blow your mind and inspire you to no end! Just three years ago, Bailey opened her Etsy shop selling PNG designs for sublimation tumblers—and since then, she's become a top 0.1% Etsy Seller earning $45k per month passively through digital downloads. Listen in as she shares her story, key strategies that helped her grow, and how she's scaling and securing her business for the future. **“How to Sell Your Stuff on Etsy” is not affiliated with or endorsed by Etsy.com STUFF I MENTIONED: Bulk listing and editing tool Bailey mentioned (Vela): https://welcome.getvela.com/ How to Create Your Own Font in Midjourney: (Go to the Fonts and Calligraphy Chapters of the video): https://youtu.be/OY427QKSXcM?si=MlQRbJmRMSVc1QIX   Bailey's Digitally Purposed Community and Training: https://www.digitallypurposed.com/digitallypurposedsales-7407?am_id=lizzie295 Find Bailey: YouTube: https://bit.ly/BaileyYouTube   Digitally Purposed: https://bit.ly/digitallypurposed   WHAT'S HAPPENING NOW: ⭐The A.I. Print on Demand LIVE Workshop just happened on 1/17/24! Get a copy of the recording, prompts for your own mockups + POD designs, tutorials for Midjourney, Ideogram, and Dalle-3 and MORE (Use code POD50 to save $50): https://www.howtosellyourstuff.com/ai-POD-workshop-enrollment ⭐ Get my free list of 100 in Demand Micro-Niche Keywords: https://www.howtosellyourstuff.com/Micro-Niche-Demand  ⭐ Join the $4.99 monthly subscription to the In Demand Micro-Niche Keywords List: https://www.howtosellyourstuff.com/micro-niche-member  (You'll get ongoing access to the ever growing of list of hundreds--soon to be thousands-- of micro-niche opportunities on Etsy!)  ---------------------------------------------   ⭐Book a one-on-one Etsy coaching session with Lizzie: https://www.howtosellyourstuff.com/coaching ⭐Apply to be a Podcast Guest: https://bit.ly/48hFD8X Find me on Instagram and TikTok @HowtoSellYourStuff   FREE ETSY MASTERCLASS: https://www.howtosellyourstuff.com/masterclass FREE PDF DOWNLOAD: “4 Strategies I Used to Grow My Etsy Shop from $25 to $6000k/month”: https://www.howtosellyourstuff.com/site/4-strategies-opt-in   Grab my UPDATED Etsy Course for physical product sellers: “Listings that Sell 2.0” and learn how to skyrocket your Etsy business: https://www.howtosellyourstuff.com/etsy-listings-that-sell    ----- HOW TO SELL YOUR STUFF WEBSITE: https://www.howtosellyourstuff.com/ HOW TO SELL YOUR STUFF INSTAGRAM: https://www.instagram.com/howtosellyourstuff/ HOW TO SELL YOUR STUFF SHOWNOTES: https://www.howtosellyourstuff.com/blog/112 ------- THIS EPISODE IS SPONSORED BY: My Resources Me! And the Resources section on my website. If you have questions specific to your personal niche on Etsy, you should definitely come check out my Resource page at www.HowtoSellYourStuff.com/Resources where I will connect you with my favorite free and paid resources created by experts I have personally vetted. Whether you sell POD, digital products, printable, physical products and more—there's info waiting for you to help you on your Etsy journey!  Recommended Resources: https://www.howtosellyourstuff.com/resources AND Paige Hulse Law and the Creative Law Shop Whether you're just getting started on Etsy or you've been selling for years but never quite got around to the legal setup, I want to make sure you know about Attorney Paige Hulse and her Creative Law Shop. If you need legal assistance for your Etsy shop, want to register a trademark, or are looking for help with forming a business—contact Paige at https://paigehulse.com/ AND If you're looking for a well-crafted legal document that is tailored to creative entrepreneurship, but don't want the cash outlay of hiring an attorney by the hour, you can get everything you need from an LLC operating agreement, multi-person LLC agreements for partnerships, special provisions for your Etsy Shop Policies, affiliate agreements, influencer contracts, photography releases, and so much more. There are over 80 contracts available plus free resources and educational tools waiting for you at https://www.shopcreativelaw.com/ Make sure you use the code smiley10 for 10% off of anything from the Creative Law Shop! *Some of the links above are affiliate links which means I'll receive a commission if you purchase through my link, at no extra cost to you. You can see my affiliate disclosure here: https://www.howtosellyourstuff.com/affiliate-disclosure

The Sam Altman Podcast
Typography Brilliance: Ideogram's AI Image Generator Unveiled

The Sam Altman Podcast

Play Episode Listen Later Jan 10, 2024 7:13


In this episode, we explore Ideogram's latest innovation—an AI-driven image generator boasting excellent custom typography, revolutionizing the visual content creation landscape. Invest in AI Box: https://Republic.com/ai-box Get on the AI Box Waitlist: https://AIBox.ai/ AI Facebook Community

The Sam Altman Podcast
Typography Brilliance: Ideogram's AI Image Generator Unveiled

The Sam Altman Podcast

Play Episode Listen Later Jan 10, 2024 6:18


In this episode, we explore Ideogram's latest innovation—an AI-driven image generator boasting excellent custom typography, revolutionizing the visual content creation landscape. Invest in AI Box: https://Republic.com/ai-box Get on the AI Box Waitlist: https://AIBox.ai/ AI Facebook Community

The Mark Cuban Podcast
Ideogram's Breakthrough: AI Image Generator Featuring Custom Typography

The Mark Cuban Podcast

Play Episode Listen Later Jan 2, 2024 6:18


In this episode, we delve into Ideogram's groundbreaking AI image generator, focusing on the incorporation of customized typography to elevate the quality and aesthetics of generated images. Invest in AI Box: https://Republic.com/ai-box Get on the AI Box Waitlist: ⁠⁠https://AIBox.ai/⁠⁠ AI Facebook Community Learn more about AI in Video Learn more about Open AI

Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and all things Software 3.0

We are running an end of year survey for our listeners! Please let us know any feedback you have, what episodes resonated with you, and guest requests for 2024! Survey link here!Before language models became all the rage in November 2022, image generation was the hottest space in AI (it was the subject of our first piece on Latent Space!) In our interview with Sharif Shameem from Lexica we talked through the launch of StableDiffusion and the early days of that space. At the time, the toolkit was still pretty rudimentary: Lexica made it easy to search images, you had the AUTOMATIC1111 Web UI to generate locally, some HuggingFace spaces that offered inference, and eventually DALL-E 2 through OpenAI's platform, but not much beyond basic text-to-image workflows.Today's guest, Suhail Doshi, is trying to solve this with Playground AI, an image editor reimagined with AI in mind. Some of the differences compared to traditional text-to-image workflows:* Real-time preview rendering using consistency: as you change your prompt, you can see changes in real-time before doing a final rendering of it.* Style filtering: rather than having to prompt exactly how you'd like an image to look, you can pick from a whole range of filters both from Playground's model as well as Stable Diffusion (like RealVis, Starlight XL, etc). We talk about this at 25:46 in the podcast.* Expand prompt: similar to DALL-E3, Playground will do some prompt tuning for you to get better results in generation. Unlike DALL-E3, you can turn this off at any time if you are a prompting wizard* Image editing: after generation, you have tools like a magic eraser, inpainting pencil, etc. This makes it easier to do a full workflow in Playground rather than switching to another tool like Photoshop.Outside of the product, they have also trained a new model from scratch, Playground v2, which is fully open source and open weights and allows for commercial usage. They benchmarked the model against SDXL across 1,000 prompts and found that humans preferred the Playground generation 70% of the time. They had similar results on PartiPrompts:They also created a new benchmark, MJHQ-30K, for “aesthetic quality”:We introduce a new benchmark, MJHQ-30K, for automatic evaluation of a model's aesthetic quality. The benchmark computes FID on a high-quality dataset to gauge aesthetic quality.We curate the high-quality dataset from Midjourney with 10 common categories, each category with 3K samples. Following common practice, we use aesthetic score and CLIP score to ensure high image quality and high image-text alignment. Furthermore, we take extra care to make the data diverse within each category.Suhail was pretty open with saying that Midjourney is currently the best product for imagine generation out there, and that's why they used it as the base for this benchmark. I think it's worth comparing yourself to maybe the best thing and try to find like a really fair way of doing that. So I think more people should try to do that. I definitely don't think you should be kind of comparing yourself on like some Google model or some old SD, Stable Diffusion model and be like, look, we beat Stable Diffusion 1.5. I think users ultimately want care, how close are you getting to the thing that people mostly agree with? [00:23:47]We also talked a lot about Suhail's founder journey from starting Mixpanel in 2009, then going through YC again with Mighty, and eventually sunsetting that to pivot into Playground. Enjoy!Show Notes* Suhail's Twitter* “Starting my road to learn AI”* Bill Gates book trip* Playground* Playground v2 Announcement* $40M raise announcement* “Running infra dev ops for 24 A100s”* Mixpanel* Mighty* “I decided to stop working on Mighty”* Fast.ai* CivitTimestamps* [00:00:00] Intros* [00:02:59] Being early in ML at Mixpanel* [00:04:16] Pivoting from Mighty to Playground and focusing on generative AI* [00:07:54] How DALL-E 2 inspired Mighty* [00:09:19] Reimagining the graphics editor with AI* [00:17:34] Training the Playground V2 model from scratch to advance generative graphics* [00:21:11] Techniques used to improve Playground V2 like data filtering and model tuning* [00:25:21] Releasing the MJHQ30K benchmark to evaluate generative models* [00:30:35] The limitations of current models for detailed image editing tasks* [00:34:06] Using post-generation user feedback to create better benchmarks* [00:38:28] Concerns over potential misuse of powerful generative models* [00:41:54] Rethinking the graphics editor user experience in the AI era* [00:45:44] Integrating consistency models into Playground using preview rendering* [00:47:23] Interacting with the Stable Diffusion LoRAs community* [00:51:35] Running DevOps on A100s* [00:53:12] Startup ideas?TranscriptAlessio: Hey everyone, welcome to the Latent Space podcast. This is Alessio, partner and CTO-in-Residence at Decibel Partners, and I'm joined by my co-host Swyx, founder of Smol AI. [00:00:15]Swyx: Hey, and today in the studio we have Suhail Doshi, welcome. [00:00:18]Suhail: Yeah, thanks. Thanks for having me. [00:00:20]Swyx: So among many things, you're a CEO and co-founder of Mixpanel, and I think about three years ago you left to start Mighty, and more recently, I think about a year ago, transitioned into Playground, and you've just announced your new round. How do you like to be introduced beyond that? [00:00:34]Suhail: Just founder of Playground is fine, yeah, prior co-founder and CEO of Mixpanel. [00:00:40]Swyx: Yeah, awesome. I'd just like to touch on Mixpanel a little bit, because it's obviously one of the more successful analytics companies we previously had amplitude on, and I'm curious if you had any reflections on the interaction of that amount of data that people would want to use for AI. I don't know if there's still a part of you that stays in touch with that world. [00:00:59]Suhail: Yeah, I mean, the short version is that maybe back in like 2015 or 2016, I don't really remember exactly, because it was a while ago, we had an ML team at Mixpanel, and I think this is when maybe deep learning or something really just started getting kind of exciting, and we were thinking that maybe given that we had such vast amounts of data, perhaps we could predict things. So we built two or three different features, I think we built a feature where we could predict whether users would churn from your product. We made a feature that could predict whether users would convert, we built a feature that could do anomaly detection, like if something occurred in your product, that was just very surprising, maybe a spike in traffic in a particular region, can we tell you that that happened? Because it's really hard to like know everything that's going on with your data, can we tell you something surprising about your data? And we tried all of these various features, most of it boiled down to just like, you know, using logistic regression, and it never quite seemed very groundbreaking in the end. And so I think, you know, we had a four or five person ML team, and I think we never expanded it from there. And I did all these Fast AI courses trying to learn about ML. And that was the- That's the first time you did fast AI. Yeah, that was the first time I did fast AI. Yeah, I think I've done it now three times, maybe. [00:02:12]Swyx: Oh, okay. [00:02:13]Suhail: I didn't know it was the third. No, no, just me reviewing it, it's maybe three times, but yeah. [00:02:16]Swyx: You mentioned prediction, but honestly, like it's also just about the feedback, right? The quality of feedback from users, I think it's useful for anyone building AI applications. [00:02:25]Suhail: Yeah. Yeah, I think I haven't spent a lot of time thinking about Mixpanel because it's been a long time, but sometimes I'm like, oh, I wonder what we could do now. And then I kind of like move on to whatever I'm working on, but things have changed significantly since. [00:02:39]Swyx: And then maybe we'll touch on Mighty a little bit. Mighty was very, very bold. My framing of it was, you will run our browsers for us because everyone has too many tabs open. I have too many tabs open and slowing down your machines that you can do it better for us in a centralized data center. [00:02:51]Suhail: Yeah, we were first trying to make a browser that we would stream from a data center to your computer at extremely low latency, but the real objective wasn't trying to make a browser or anything like that. The real objective was to try to make a new kind of computer. And the thought was just that like, you know, we have these computers in front of us today and we upgrade them or they run out of RAM or they don't have enough RAM or not enough disk or, you know, there's some limitation with our computers, perhaps like data locality is a problem. Why do I need to think about upgrading my computer ever? And so, you know, we just had to kind of observe that like, well, actually it seems like a lot of applications are just now in the browser, you know, it's like how many real desktop applications do we use relative to the number of applications we use in the browser? So it's just this realization that actually like, you know, the browser was effectively becoming more or less our operating system over time. And so then that's why we kind of decided to go, hmm, maybe we can stream the browser. Fortunately, the idea did not work for a couple of different reasons, but the objective is try to make sure new computer. [00:03:50]Swyx: Yeah, very, very bold. [00:03:51]Alessio: Yeah, and I was there at YC Demo Day when you first announced it. It was, I think, the last or one of the last in-person ones, at Pier34 in Mission Bay. How do you think about that now when everybody wants to put some of these models in people's machines and some of them want to stream them in, do you think there's maybe another wave of the same problem before it was like browser apps too slow, now it's like models too slow to run on device? [00:04:16]Suhail: Yeah. I mean, I've obviously pivoted away from Mighty, but a lot of what I somewhat believed at Mighty, maybe why I'm so excited about AI and what's happening, a lot of what Mighty was about was like moving compute somewhere else, right? Right now, applications, they get limited quantities of memory, disk, networking, whatever your home network has, et cetera. You know, what if these applications could somehow, if we could shift compute, and then these applications have vastly more compute than they do today. Right now it's just like client backend services, but you know, what if we could change the shape of how applications could interact with things? And it's changed my thinking. In some ways, AI has like a bit of a continuation of my belief that like perhaps we can really shift compute somewhere else. One of the problems with Mighty was that JavaScript is single-threaded in the browser. And what we learned, you know, the reason why we kind of abandoned Mighty was because I didn't believe we could make a new kind of computer. We could have made some kind of enterprise business, probably it could have made maybe a lot of money, but it wasn't going to be what I hoped it was going to be. And so once I realized that most of a web app is just going to be single-threaded JavaScript, then the only thing you could do largely withstanding changing JavaScript, which is a fool's errand most likely, make a better CPU, right? And there's like three CPU manufacturers, two of which sell, you know, big ones, you know, AMD, Intel, and then of course like Apple made the M1. And it's not like single-threaded CPU core performance, single-core performance was increasing very fast, it's plateauing rapidly. And even these different companies were not doing as good of a job, you know, sort of with the continuation of Moore's law. But what happened in AI was that you got like, if you think of the AI model as like a computer program, like just like a compiled computer program, it is literally built and designed to do massive parallel computations. And so if you could take like the universal approximation theorem to its like kind of logical complete point, you know, you're like, wow, I can get, make computation happen really rapidly and parallel somewhere else, you know, so you end up with these like really amazing models that can like do anything. It just turned out like perhaps the new kind of computer would just simply be shifted, you know, into these like really amazing AI models in reality. Yeah. [00:06:30]Swyx: Like I think Andrej Karpathy has always been, has been making a lot of analogies with the LLMOS. [00:06:34]Suhail: I saw his video and I watched that, you know, maybe two weeks ago or something like that. I was like, oh man, this, I very much resonate with this like idea. [00:06:41]Swyx: Why didn't I see this three years ago? [00:06:43]Suhail: Yeah. I think, I think there still will be, you know, local models and then there'll be these very large models that have to be run in data centers. I think it just depends on kind of like the right tool for the job, like any engineer would probably care about. But I think that, you know, by and large, like if the models continue to kind of keep getting bigger, you're always going to be wondering whether you should use the big thing or the small, you know, the tiny little model. And it might just depend on like, you know, do you need 30 FPS or 60 FPS? Maybe that would be hard to do, you know, over a network. [00:07:13]Swyx: You tackled a much harder problem latency wise than the AI models actually require. Yeah. [00:07:18]Suhail: Yeah. You can do quite well. You can do quite well. You definitely did 30 FPS video streaming, did very crazy things to make that work. So I'm actually quite bullish on the kinds of things you can do with networking. [00:07:30]Swyx: Maybe someday you'll come back to that at some point. But so for those that don't know, you're very transparent on Twitter. Very good to follow you just to learn your insights. And you actually published a postmortem on Mighty that people can read up on and willing to. So there was a bit of an overlap. You started exploring the AI stuff in June 2022, which is when you started saying like, I'm taking fast AI again. Maybe, was there more context around that? [00:07:54]Suhail: Yeah. I think I was kind of like waiting for the team at Mighty to finish up, you know, something. And I was like, okay, well, what can I do? I guess I will make some kind of like address bar predictor in the browser. So we had, you know, we had forked Chrome and Chromium. And I was like, you know, one thing that's kind of lame is that like this browser should be like a lot better at predicting what I might do, where I might want to go. It struck me as really odd that, you know, Chrome had very little AI actually or ML inside this browser. For a company like Google, you'd think there's a lot. Code is actually just very, you know, it's just a bunch of if then statements is more or less the address bar. So it seemed like a pretty big opportunity. And that's also where a lot of people interact with the browser. So, you know, long story short, I was like, hmm, I wonder what I could build here. So I started to take some AI courses and review the material again and get back to figuring it out. But I think that was somewhat serendipitous because right around April was, I think, a very big watershed moment in AI because that's when Dolly 2 came out. And I think that was the first truly big viral moment for generative AI. [00:08:59]Swyx: Because of the avocado chair. [00:09:01]Suhail: Yeah, exactly. [00:09:02]Swyx: It wasn't as big for me as Stable Diffusion. [00:09:04]Suhail: Really? [00:09:05]Swyx: Yeah, I don't know. Dolly was like, all right, that's cool. [00:09:07]Suhail: I don't know. Yeah. [00:09:09]Swyx: I mean, they had some flashy videos, but it didn't really register. [00:09:13]Suhail: That moment of images was just such a viral novel moment. I think it just blew people's mind. Yeah. [00:09:19]Swyx: I mean, it's the first time I encountered Sam Altman because they had this Dolly 2 hackathon and they opened up the OpenAI office for developers to walk in back when it wasn't as much of a security issue as it is today. I see. Maybe take us through the journey to decide to pivot into this and also choosing images. Obviously, you were inspired by Dolly, but there could be any number of AI companies and businesses that you could start and why this one, right? [00:09:45]Suhail: Yeah. So I think at that time, Mighty and OpenAI was not quite as popular as it is all of a sudden now these days, but back then they had a lot more bandwidth to kind of help anybody. And so we had been talking with the team there around trying to see if we could do really fast low latency address bar prediction with GPT-3 and 3.5 and that kind of thing. And so we were sort of figuring out how could we make that low latency. I think that just being able to talk to them and kind of being involved gave me a bird's eye view into a bunch of things that started to happen. Latency first was the Dolly 2 moment, but then stable diffusion came out and that was a big moment for me as well. And I remember just kind of like sitting up one night thinking, I was like, you know, what are the kinds of companies one could build? Like what matters right now? One thing that I observed is that I find a lot of inspiration when I'm working in a field in something and then I can identify a bunch of problems. Like for Mixpanel, I was an intern at a company and I just noticed that they were doing all this data analysis. And so I thought, hmm, I wonder if I could make a product and then maybe they would use it. And in this case, you know, the same thing kind of occurred. It was like, okay, there are a bunch of like infrastructure companies that put a model up and then you can use their API, like Replicate is a really good example of that. There are a bunch of companies that are like helping you with training, model optimization, Mosaic at the time, and probably still, you know, was doing stuff like that. So I just started listing out like every category of everything, of every company that was doing something interesting. I started listing out like weights and biases. I was like, oh man, weights and biases is like this great company. Do I want to compete with that company? I might be really good at competing with that company because of Mixpanel because it's so much of like analysis. But I was like, no, I don't want to do anything related to that. That would, I think that would be too boring now at this point. So I started to list out all these ideas and one thing I observed was that at OpenAI, they had like a playground for GPT-3, right? All it was is just like a text box more or less. And then there were some settings on the right, like temperature and whatever. [00:11:41]Swyx: Top K. [00:11:42]Suhail: Yeah, top K. You know, what's your end stop sequence? I mean, that was like their product before GPT, you know, really difficult to use, but fun if you're like an engineer. And I just noticed that their product kind of was evolving a little bit where the interface kind of was getting a little bit more complex. They had like a way where you could like generate something in the middle of a sentence and all those kinds of things. And I just thought to myself, I was like, everything is just like this text box and you generate something and that's about it. And stable diffusion had kind of come out and it was all like hugging face and code. Nobody was really building any UI. And so I had this kind of thing where I wrote prompt dash like question mark in my notes and I didn't know what was like the product for that at the time. I mean, it seems kind of trite now, but I just like wrote prompt. What's the thing for that? Manager. Prompt manager. Do you organize them? Like, do you like have a UI that can play with them? Yeah. Like a library. What would you make? And so then, of course, then you thought about what would the modalities be given that? How would you build a UI for each kind of modality? And so there are a couple of people working on some pretty cool things. And I basically chose graphics because it seemed like the most obvious place where you could build a really powerful, complex UI. That's not just only typing a box. It would very much evolve beyond that. Like what would be the best thing for something that's visual? Probably something visual. Yeah. I think that just that progression kind of happened and it just seemed like there was a lot of effort going into language, but not a lot of effort going into graphics. And then maybe the very last thing was, I think I was talking to Aditya Ramesh, who was the co-creator of DALL-E 2 and Sam. And I just kind of went to these guys and I was just like, hey, are you going to make like a UI for this thing? Like a true UI? Are you going to go for this? Are you going to make a product? For DALL-E. Yeah. For DALL-E. Yeah. Are you going to do anything here? Because if you are going to do it, just let me know and I will stop and I'll go do something else. But if you're not going to do anything, I'll just do it. And so we had a couple of conversations around what that would look like. And then I think ultimately they decided that they were going to focus on language primarily. And I just felt like it was going to be very underinvested in. Yes. [00:13:46]Swyx: There's that sort of underinvestment from OpenAI, but also it's a different type of customer than you're used to, presumably, you know, and Mixpanel is very good at selling to B2B and developers will figure on you or not. Yeah. Was that not a concern? [00:14:00]Suhail: Well, not so much because I think that, you know, right now I would say graphics is in this very nascent phase. Like most of the customers are just like hobbyists, right? Yeah. Like it's a little bit of like a novel toy as opposed to being this like very high utility thing. But I think ultimately, if you believe that you could make it very high utility, the probably the next customers will end up being B2B. It'll probably not be like a consumer. There will certainly be a variation of this idea that's in consumer. But if your quest is to kind of make like something that surpasses human ability for graphics, like ultimately it will end up being used for business. So I think it's maybe more of a progression. In fact, for me, it's maybe more like Mixpanel started out as SMB and then very much like ended up starting to grow up towards enterprise. So for me, I think it will be a very similar progression. But yeah, I mean, the reason why I was excited about it is because it was a creative tool. I make music and it's AI. It's like something that I know I could stay up till three o'clock in the morning doing. Those are kind of like very simple bars for me. [00:14:56]Alessio: So you mentioned Dolly, Stable Diffusion. You just had Playground V2 come out two days ago. Yeah, two days ago. [00:15:02]Suhail: Two days ago. [00:15:03]Alessio: This is a model you train completely from scratch. So it's not a cheap fine tune on something. You open source everything, including the weights. Why did you decide to do it? I know you supported Stable Diffusion XL in Playground before, right? Yep. What made you want to come up with V2 and maybe some of the interesting, you know, technical research work you've done? [00:15:24]Suhail: Yeah. So I think that we continue to feel like graphics and these foundation models for anything really related to pixels, but also definitely images continues to be very underinvested. It feels a little like graphics is in like this GPT-2 moment, right? Like even GPT-3, even when GPT-3 came out, it was exciting, but it was like, what are you going to use this for? Yeah, we'll do some text classification and some semantic analysis and maybe it'll sometimes like make a summary of something and it'll hallucinate. But no one really had like a very significant like business application for GPT-3. And in images, we're kind of stuck in the same place. We're kind of like, okay, I write this thing in a box and I get some cool piece of artwork and the hands are kind of messed up and sometimes the eyes are a little weird. Maybe I'll use it for a blog post, you know, that kind of thing. The utility feels so limited. And so, you know, and then we, you sort of look at Stable Diffusion and we definitely use that model in our product and our users like it and use it and love it and enjoy it, but it hasn't gone nearly far enough. So we were kind of faced with the choice of, you know, do we wait for progress to occur or do we make that progress happen? So yeah, we kind of embarked on a plan to just decide to go train these things from scratch. And I think the community has given us so much. The community for Stable Diffusion I think is one of the most vibrant communities on the internet. It's like amazing. It feels like, I hope this is what like Homebrew Club felt like when computers like showed up because it's like amazing what that community will do and it moves so fast. I've never seen anything in my life and heard other people's stories around this where an academic research paper comes out and then like two days later, someone has sample code for it. And then two days later, there's a model. And then two days later, it's like in nine products, you know, they're all competing with each other. It's incredible to see like math symbols on an academic paper go to well-designed features in a product. So I think the community has done so much. So I think we wanted to give back to the community kind of on our way. Certainly we would train a better model than what we gave out on Tuesday, but we definitely felt like there needs to be some kind of progress in these open source models. The last kind of milestone was in July when Stable Diffusion Excel came out, but there hasn't been anything really since. Right. [00:17:34]Swyx: And there's Excel Turbo now. [00:17:35]Suhail: Well, Excel Turbo is like this distilled model, right? So it's like lower quality, but fast. You have to decide, you know, what your trade off is there. [00:17:42]Swyx: It's also a consistency model. [00:17:43]Suhail: I don't think it's a consistency model. It's like it's they did like a different thing. Yeah. I think it's like, I don't want to get quoted for this, but it's like something called ad like adversarial or something. [00:17:52]Swyx: That's exactly right. [00:17:53]Suhail: I've read something about that. Maybe it's like closer to GANs or something, but I didn't really read the full paper. But yeah, there hasn't been quite enough progress in terms of, you know, there's no multitask image model. You know, the closest thing would be something called like EmuEdit, but there's no model for that. It's just a paper that's within meta. So we did that and we also gave out pre-trained weights, which is very rare. Usually you just get the aligned model and then you have to like see if you can do anything with it. So we actually gave out, there's like a 256 pixel pre-trained stage and a 512. And we did that for academic research because we come across people all the time in academia, they have access to like one A100 or eight at best. And so if we can give them kind of like a 512 pre-trained model, our hope is that there'll be interesting novel research that occurs from that. [00:18:38]Swyx: What research do you want to happen? [00:18:39]Suhail: I would love to see more research around things that users care about tend to be things like character consistency. [00:18:45]Swyx: Between frames? [00:18:46]Suhail: More like if you have like a face. Yeah, yeah. Basically between frames, but more just like, you know, you have your face and it's in one image and then you want it to be like in another. And users are very particular and sensitive to faces changing because we know we're trained on faces as humans. Not seeing a lot of innovation, enough innovation around multitask editing. You know, there are two things like instruct pics to pics and then the EmuEdit paper that are maybe very interesting, but we certainly are not pushing the fold on that in that regard. All kinds of things like around that rotation, you know, being able to keep coherence across images, style transfer is still very limited. Just even reasoning around images, you know, what's going on in an image, that kind of thing. Things are still very, very underpowered, very nascent. So therefore the utility is very, very limited. [00:19:32]Alessio: On the 1K Prompt Benchmark, you are 2.5x prefer to Stable Diffusion XL. How do you get there? Is it better images in the training corpus? Can you maybe talk through the improvements in the model? [00:19:44]Suhail: I think they're still very early on in the recipe, but I think it's a lot of like little things and you know, every now and then there are some big important things like certainly your data quality is really, really important. So we spend a lot of time thinking about that. But I would say it's a lot of things that you kind of clean up along the way as you train your model. Everything from captions to the data that you align with after pre-train to how you're picking your data sets, how you filter your data sets. I feel like there's a lot of work in AI that doesn't really feel like AI. It just really feels like just data set filtering and systems engineering and just like, you know, and the recipe is all there, but it's like a lot of extra work to do that. I think we plan to do a Playground V 2.1, maybe either by the end of the year or early next year. And we're just like watching what the community does with the model. And then we're just going to take a lot of the things that they're unhappy about and just like fix them. You know, so for example, like maybe the eyes of people in an image don't feel right. They feel like they're a little misshapen or they're kind of blurry feeling. That's something that we already know we want to fix. So I think in that case, it's going to be about data quality. Or maybe you want to improve the kind of the dynamic range of color. You know, we want to make sure that that's like got a good range in any image. So what technique can we use there? There's different things like offset noise, pyramid noise, terminal zero, SNR, like there are all these various interesting things that you can do. So I think it's like a lot of just like tricks. Some are tricks, some are data, and some is just like cleaning. [00:21:11]Swyx: Specifically for faces, it's very common to use a pipeline rather than just train the base model more. Do you have a strong belief either way on like, oh, they should be separated out to different stages for like improving the eyes, improving the face or enhance or whatever? Or do you think like it can all be done in one model? [00:21:28]Suhail: I think we will make a unified model. Yeah, I think it will. I think we'll certainly in the end, ultimately make a unified model. There's not enough research about this. Maybe there is something out there that we haven't read. There are some bottlenecks, like for example, in the VAE, like the VAEs are ultimately like compressing these things. And so you don't know. And then you might have like a big informational information bottleneck. So maybe you would use a pixel based model, perhaps. I think we've talked to people, everyone from like Rombach to various people, Rombach trained stable diffusion. I think there's like a big question around the architecture of these things. It's still kind of unknown, right? Like we've got transformers and we've got like a GPT architecture model, but then there's this like weird thing that's also seemingly working with diffusion. And so, you know, are we going to use vision transformers? Are we going to move to pixel based models? Is there a different kind of architecture? We don't really, I don't think there have been enough experiments. Still? Oh my God. [00:22:21]Swyx: Yeah. [00:22:22]Suhail: That's surprising. I think it's very computationally expensive to do a pipeline model where you're like fixing the eyes and you're fixing the mouth and you're fixing the hands. [00:22:29]Swyx: That's what everyone does as far as I understand. [00:22:31]Suhail: I'm not exactly sure what you mean, but if you mean like you get an image and then you will like make another model specifically to fix a face, that's fairly computationally expensive. And I think it's like not probably not the right way. Yeah. And it doesn't generalize very well. Now you have to pick all these different things. [00:22:45]Swyx: Yeah. You're just kind of glomming things on together. Yeah. Like when I look at AI artists, like that's what they do. [00:22:50]Suhail: Ah, yeah, yeah, yeah. They'll do things like, you know, I think a lot of ARs will do control net tiling to do kind of generative upscaling of all these different pieces of the image. Yeah. And I think these are all just like, they're all hacks ultimately in the end. I mean, it just to me, it's like, let's go back to where we were just three years, four years ago with where deep learning was at and where language was that, you know, it's the same thing. It's like we were like, okay, well, I'll just train these very narrow models to try to do these things and kind of ensemble them or pipeline them to try to get to a best in class result. And here we are with like where the models are gigantic and like very capable of solving huge amounts of tasks when given like lots of great data. [00:23:28]Alessio: You also released a new benchmark called MJHQ30K for automatic evaluation of a model's aesthetic quality. I have one question. The data set that you use for the benchmark is from Midjourney. Yes. You have 10 categories. How do you think about the Playground model, Midjourney, like, are you competitors? [00:23:47]Suhail: There are a lot of people, a lot of people in research, they like to compare themselves to something they know they can beat, right? Maybe this is the best reason why it can be helpful to not be a researcher also sometimes like I'm not trained as a researcher, I don't have a PhD in anything AI related, for example. But I think if you care about products and you care about your users, then the most important thing that you want to figure out is like everyone has to acknowledge that Midjourney is very good. They are the best at this thing. I'm happy to admit that. I have no problem admitting that. Just easy. It's very visual to tell. So I think it's incumbent on us to try to compare ourselves to the thing that's best, even if we lose, even if we're not the best. At some point, if we are able to surpass Midjourney, then we only have ourselves to compare ourselves to. But on First Blush, I think it's worth comparing yourself to maybe the best thing and try to find like a really fair way of doing that. So I think more people should try to do that. I definitely don't think you should be kind of comparing yourself on like some Google model or some old SD, Stable Diffusion model and be like, look, we beat Stable Diffusion 1.5. I think users ultimately want care, how close are you getting to the thing that people mostly agree with? So we put out that benchmark for no other reason to say like, this seems like a worthy thing for us to at least try, for people to try to get to. And then if we surpass it, great, we'll come up with another one. [00:25:06]Alessio: Yeah, no, that's awesome. And you killed Stable Diffusion Excel and everything. In the benchmark chart, it says Playground V2 1024 pixel dash aesthetic. Do you have kind of like, yeah, style fine tunes or like what's the dash aesthetic for? [00:25:21]Suhail: We debated this, maybe we named it wrong or something, but we were like, how do we help people realize the model that's aligned versus the models that weren't? Because we gave out pre-trained models, we didn't want people to like use those. So that's why they're called base. And then the aesthetic model, yeah, we wanted people to pick up the thing that makes things pretty. Who wouldn't want the thing that's aesthetic? But if there's a better name, we're definitely open to feedback. No, no, that's cool. [00:25:46]Alessio: I was using the product. You also have the style filter and you have all these different styles. And it seems like the styles are tied to the model. So there's some like SDXL styles, there's some Playground V2 styles. Can you maybe give listeners an overview of how that works? Because in language, there's not this idea of like style, right? Versus like in vision model, there is, and you cannot get certain styles in different [00:26:11]Suhail: models. [00:26:12]Alessio: So how do styles emerge and how do you categorize them and find them? [00:26:15]Suhail: Yeah, I mean, it's so fun having a community where people are just trying a model. Like it's only been two days for Playground V2. And we actually don't know what the model's capable of and not capable of. You know, we certainly see problems with it. But we have yet to see what emergent behavior is. I mean, we've just sort of discovered that it takes about like a week before you start to see like new things. I think like a lot of that style kind of emerges after that week, where you start to see, you know, there's some styles that are very like well known to us, like maybe like pixel art is a well known style. Photorealism is like another one that's like well known to us. But there are some styles that cannot be easily named. You know, it's not as simple as like, okay, that's an anime style. It's very visual. And in the end, you end up making up the name for what that style represents. And so the community kind of shapes itself around these different things. And so if anyone that's into stable diffusion and into building anything with graphics and stuff with these models, you know, you might have heard of like Proto Vision or Dream Shaper, some of these weird names, but they're just invented by these authors. But they have a sort of je ne sais quoi that, you know, appeals to users. [00:27:26]Swyx: Because it like roughly embeds to what you what you want. [00:27:29]Suhail: I guess so. I mean, it's like, you know, there's one of my favorite ones that's fine tuned. It's not made by us. It's called like Starlight XL. It's just this beautiful model. It's got really great color contrast and visual elements. And the users love it. I love it. And it's so hard. I think that's like a very big open question with graphics that I'm not totally sure how we'll solve. I don't know. It's, it's like an evolving situation too, because styles get boring, right? They get fatigued. Like it's like listening to the same style of pop song. I try to relate to graphics a little bit like with music, because I think it gives you a little bit of a different shape to things. Like it's not as if we just have pop music, rap music and country music, like all of these, like the EDM genre alone has like sub genres. And I think that's very true in graphics and painting and art and anything that we're doing. There's just these sub genres, even if we can't quite always name them. But I think they are emergent from the community, which is why we're so always happy to work with the community. [00:28:26]Swyx: That is a struggle. You know, coming back to this, like B2B versus B2C thing, B2C, you're going to have a huge amount of diversity and then it's going to reduce as you get towards more sort of B2B type use cases. I'm making this up here. So like you might be optimizing for a thing that you may eventually not need. [00:28:42]Suhail: Yeah, possibly. Yeah, possibly. I think like a simple thing with startups is that I worry sometimes by trying to be overly ambitious and like really scrutinizing like what something is in its most nascent phase that you miss the most ambitious thing you could have done. Like just having like very basic curiosity with something very small can like kind of lead you to something amazing. Like Einstein definitely did that. And then he like, you know, he basically won all the prizes and got everything he wanted and then basically did like kind of didn't really. He can dismiss quantum and then just kind of was still searching, you know, for the unifying theory. And he like had this quest. I think that happens a lot with like Nobel Prize people. I think there's like a term for it that I forget. I actually wanted to go after a toy almost intentionally so long as that I could see, I could imagine that it would lead to something very, very large later. Like I said, it's very hobbyist, but you need to start somewhere. You need to start with something that has a big gravitational pull, even if these hobbyists aren't likely to be the people that, you know, have a way to monetize it or whatever, even if they're, but they're doing it for fun. So there's something, something there that I think is really important. But I agree with you that, you know, in time we will absolutely focus on more utilitarian things like things that are more related to editing feats that are much harder. And so I think like a very simple use case is just, you know, I'm not a graphics designer. It seems like very simple that like you, if we could give you the ability to do really complex graphics without skill, wouldn't you want that? You know, like my wife the other day was set, you know, said, I wish Playground was better. When are you guys going to have a feature where like we could make my son, his name's Devin, smile when he was not smiling in the picture for the holiday card. Right. You know, just being able to highlight his, his mouth and just say like, make him smile. Like why can't we do that with like high fidelity and coherence, little things like that, all the way to putting you in completely different scenarios. [00:30:35]Swyx: Is that true? Can we not do that in painting? [00:30:37]Suhail: You can do in painting, but the quality is just so bad. Yeah. It's just really terrible quality. You know, it's like you'll do it five times and it'll still like kind of look like crooked or just artifact. Part of it's like, you know, the lips on the face, there's such little information there. So small that the models really struggle with it. Yeah. [00:30:55]Swyx: Make the picture smaller and you don't see it. That's my trick. I don't know. [00:30:59]Suhail: Yeah. Yeah. That's true. Or, you know, you could take that region and make it really big and then like say it's a mouth and then like shrink it. It feels like you're wrestling with it more than it's doing something that kind of surprises you. [00:31:12]Swyx: Yeah. It feels like you are very much the internal tastemaker, like you carry in your head this vision for what a good art model should look like. Do you find it hard to like communicate it to like your team and other people? Just because it's obviously it's hard to put into words like we just said. [00:31:26]Suhail: Yeah. It's very hard to explain. Images have such high bitrate compared to just words and we don't have enough words to describe these things. It's not terribly difficult. I think everyone on the team, if they don't have good kind of like judgment taste or like an eye for some of these things, they're like steadily building it because they have no choice. Right. So in that realm, I don't worry too much, actually. Like everyone is kind of like learning to get the eye is what I would call it. But I also have, you know, my own narrow taste. Like I don't represent the whole population either. [00:31:59]Swyx: When you benchmark models, you know, like this benchmark we're talking about, we use FID. Yeah. Input distance. OK. That's one measure. But like it doesn't capture anything you just said about smiles. [00:32:08]Suhail: Yeah. FID is generally a bad metric. It's good up to a point and then it kind of like is irrelevant. Yeah. [00:32:14]Swyx: And then so are there any other metrics that you like apart from vibes? I'm always looking for alternatives to vibes because vibes don't scale, you know. [00:32:22]Suhail: You know, it might be fun to kind of talk about this because it's actually kind of fresh. So up till now, we haven't needed to do a ton of like benchmarking because we hadn't trained our own model and now we have. So now what? What does that mean? How do we evaluate it? And, you know, we're kind of like living with the last 48, 72 hours of going, did the way that we benchmark actually succeed? [00:32:43]Swyx: Did it deliver? [00:32:44]Suhail: Right. You know, like I think Gemini just came out. They just put out a bunch of benchmarks. But all these benchmarks are just an approximation of how you think it's going to end up with real world performance. And I think that's like very fascinating to me. So if you fake that benchmark, you'll still end up in a really bad scenario at the end of the day. And so, you know, one of the benchmarks we did was we kind of curated like a thousand prompts. And I think that's kind of what we published in our blog post, you know, of all these tasks that we a lot of some of them are curated by our team where we know the models all suck at it. Like my favorite prompt that no model is really capable of is a horse riding an astronaut, the inverse one. And it's really, really hard to do. [00:33:22]Swyx: Not in data. [00:33:23]Suhail: You know, another one is like a giraffe underneath a microwave. How does that work? Right. There's so many of these little funny ones. We do. We have prompts that are just like misspellings of things. Yeah. We'll figure out if the models will figure it out. [00:33:36]Swyx: They should embed to the same space. [00:33:39]Suhail: Yeah. And just like all these very interesting weirdo things. And so we have so many of these and then we kind of like evaluate whether the models are any good at it. And the reality is that they're all bad at it. And so then you're just picking the most aesthetic image. We're still at the beginning of building like the best benchmark we can that aligns most with just user happiness, I think, because we're not we're not like putting these in papers and trying to like win, you know, I don't know, awards at ICCV or something if they have awards. You could. [00:34:05]Swyx: That's absolutely a valid strategy. [00:34:06]Suhail: Yeah, you could. But I don't think it could correlate necessarily with the impact we want to have on humanity. I think we're still evolving whatever our benchmarks are. So the first benchmark was just like very difficult tasks that we know the models are bad at. Can we come up with a thousand of these, whether they're hand rated and some of them are generated? And then can we ask the users, like, how do we do? And then we wanted to use a benchmark like party prompts. We mostly did that so people in academia could measure their models against ours versus others. But yeah, I mean, fit is pretty bad. And I think in terms of vibes, it's like you put out the model and then you try to see like what users make. And I think my sense is that we're going to take all the things that we notice that the users kind of were failing at and try to find like new ways to measure that, whether that's like a smile or, you know, color contrast or lighting. One benefit of Playground is that we have users making millions of images every single day. And so we can just ask them for like a post generation feedback. Yeah, we can just ask them. We can just say, like, how good was the lighting here? How was the subject? How was the background? [00:35:06]Swyx: Like a proper form of like, it's just like you make it, you come to our site, you make [00:35:10]Suhail: an image and then we say, and then maybe randomly you just say, hey, you know, like, how was the color and contrast of this image? And you say it was not very good, just tell us. So I think I think we can get like tens of thousands of these evaluations every single day to truly measure real world performance as opposed to just like benchmark performance. I would like to publish hopefully next year. I think we will try to publish a benchmark that anyone could use, that we evaluate ourselves on and that other people can, that we think does a good job of approximating real world performance because we've tried it and done it and noticed that it did. Yeah. I think we will do that. [00:35:45]Swyx: I personally have a few like categories that I consider special. You know, you know, you have like animals, art, fashion, food. There are some categories which I consider like a different tier of image. Top among them is text in images. How do you think about that? So one of the big wow moments for me, something I've been looking out for the entire year is just the progress of text and images. Like, can you write in an image? Yeah. And Ideogram came out recently, which had decent but not perfect text and images. Dolly3 had improved some and all they said in their paper was that they just included more text in the data set and it just worked. I was like, that's just lazy. But anyway, do you care about that? Because I don't see any of that in like your sample. Yeah, yeah. [00:36:27]Suhail: The V2 model was mostly focused on image quality versus like the feature of text synthesis. [00:36:33]Swyx: Well, as a business user, I care a lot about that. [00:36:35]Suhail: Yeah. Yeah. I'm very excited about text synthesis. And yeah, I think Ideogram has done a good job of maybe the best job. Dolly has like a hit rate. Yes. You know, like sometimes it's Egyptian letters. Yeah. I'm very excited about text synthesis. You know, I don't have much to say on it just yet. You know, you don't want just text effects. I think where this has to go is it has to be like you could like write little tiny pieces of text like on like a milk carton. That's maybe not even the focal point of a scene. I think that's like a very hard task that, you know, if you could do something like that, then there's a lot of other possibilities. Well, you don't have to zero shot it. [00:37:09]Swyx: You can just be like here and focus on this. [00:37:12]Suhail: Sure. Yeah, yeah. Definitely. Yeah. [00:37:16]Swyx: Yeah. So I think text synthesis would be very exciting. I'll also flag that Max Wolf, MiniMaxxier, which you must have come across his work. He's done a lot of stuff about using like logo masks that then map onto food and vegetables. And it looks like text, which can be pretty fun. [00:37:29]Suhail: That's the wonderful thing about like the open source community is that you get things like control net and then you see all these people do these just amazing things with control net. And then you wonder, I think from our point of view, we sort of go that that's really wonderful. But how do we end up with like a unified model that can do that? What are the bottlenecks? What are the issues? The community ultimately has very limited resources. And so they need these kinds of like workaround research ideas to get there. But yeah. [00:37:55]Swyx: Are techniques like control net portable to your architecture? [00:37:58]Suhail: Definitely. Yeah. We kept the Playground V2 exactly the same as SDXL. Not because not out of laziness, but just because we knew that the community already had tools. You know, all you have to do is maybe change a string in your code and then, you know, retrain a control net for it. So it was very intentional to do that. We didn't want to fragment the community with different architectures. Yeah. [00:38:16]Swyx: So basically, I'm going to go over three more categories. One is UIs, like app UIs, like mock UIs. Third is not safe for work, and then copyrighted stuff. I don't know if you care to comment on any of those. [00:38:28]Suhail: I think the NSFW kind of like safety stuff is really important. I kind of think that one of the biggest risks kind of going into maybe the U.S. election year will probably be very interrelated with like graphics, audio, video. I think it's going to be very hard to explain, you know, to a family relative who's not kind of in our world. And our world is like sometimes very, you know, we think it's very big, but it's very tiny compared to the rest of the world. Some people like there's still lots of humanity who have no idea what chat GPT is. And I think it's going to be very hard to explain, you know, to your uncle, aunt, whoever, you know, hey, I saw President Biden say this thing on a video, you know, I can't believe, you know, he said that. I think that's going to be a very troubling thing going into the world next year, the year after. [00:39:12]Swyx: That's more like a risk thing, like deepfakes, faking, political faking. But there's a lot of studies on how for most businesses, you don't want to train on not safe for work images, except that it makes you really good at bodies. [00:39:24]Suhail: Personally, we filter out NSFW type of images in our data set so that it's, you know, so our safety filter stuff doesn't have to work as hard. [00:39:32]Swyx: But you've heard this argument that not safe for work images are very good at human anatomy, which you do want to be good at. [00:39:38]Suhail: It's not like necessarily a bad thing to train on that data. It's more about like how you go and use it. That's why I was kind of talking about safety, you know, in part, because there are very terrible things that can happen in the world. If you have an extremely powerful graphics model, you know, suddenly like you can kind of imagine, you know, now if you can like generate nudes and then there's like you could do very character consistent things with faces, like what does that lead to? Yeah. And so I tend to think more what occurs after that, right? Even if you train on, let's say, you know, new data, if it does something to kind of help, there's nothing wrong with the human anatomy, it's very valid for a model to learn that. But then it's kind of like, how does that get used? And, you know, I won't bring up all of the very, very unsavory, terrible things that we see on a daily basis on the site, but I think it's more about what occurs. And so we, you know, we just recently did like a big sprint on safety. It's very difficult with graphics and art, right? Because there is tasteful art that has nudity, right? They're all over in museums, like, you know, there's very valid situations for that. And then there's the things that are the gray line of that, you know, what I might not find tasteful, someone might be like, that is completely tasteful, right? And then there are things that are way over the line. And then there are things that maybe you or, you know, maybe I would be okay with, but society isn't, you know? So where does that kind of end up on the spectrum of things? I think it's really hard with art. Sometimes even if you have like things that are not nude, if a child goes to your site, scrolls down some images, you know, classrooms of kids, you know, using our product, it's a really difficult problem. And it stretches mostly culture, society, politics, everything. [00:41:14]Alessio: Another favorite topic of our listeners is UX and AI. And I think you're probably one of the best all-inclusive editors for these things. So you don't just have the prompt, images come out, you pray, and now you do it again. First, you let people pick a seed so they can kind of have semi-repeatable generation. You also have, yeah, you can pick how many images and then you leave all of them in the canvas. And then you have kind of like this box, the generation box, and you can even cross between them and outpaint. There's all these things. How did you get here? You know, most people are kind of like, give me text, I give you image. You know, you're like, these are all the tools for you. [00:41:54]Suhail: Even though we were trying to make a graphics foundation model, I think we think that we're also trying to like re-imagine like what a graphics editor might look like given the change in technology. So, you know, I don't think we're trying to build Photoshop, but it's the only thing that we could say that people are largely familiar with. Oh, okay, there's Photoshop. What would Photoshop compare itself to pre-computer? I don't know, right? It's like, or kind of like a canvas, but you know, there's these menu options and you can use your mouse. What's a mouse? So I think that we're trying to re-imagine what a graphics editor might look like, not just for the fun of it, but because we kind of have no choice. Like there's this idea in image generation where you can generate images. That's like a super weird thing. What is that in Photoshop, right? You have to wait right now for the time being, but the wait is worth it often for a lot of people because they can't make that with their own skills. So I think it goes back to, you know, how we started the company, which was kind of looking at GPT-3's Playground, that the reason why we're named Playground is a homage to that actually. And, you know, it's like, shouldn't these products be more visual? These prompt boxes are like a terminal window, right? We're kind of at this weird point where it's just like MS-DOS. I remember my mom using MS-DOS and I memorized the keywords, like DIR, LS, all those things, right? It feels a little like we're there, right? Prompt engineering, parentheses to say beautiful or whatever, waits the word token more in the model or whatever. That's like super strange. I think a large portion of humanity would agree that that's not user-friendly, right? So how do we think about the products to be more user-friendly? Well, sure, you know, sure, it would be nice if I wanted to get rid of, like, the headphones on my head, you know, it'd be nice to mask it and then say, you know, can you remove the headphones? You know, if I want to grow, expand the image, you know, how can we make that feel easier without typing lots of words and being really confused? I don't even think we've nailed the UI UX yet. Part of that is because we're still experimenting. And part of that is because the model and the technology is going to get better. And whatever felt like the right UX six months ago is going to feel very broken now. So that's a little bit of how we got there is kind of saying, does everything have to be like a prompt in a box? Or can we do things that make it very intuitive for users? [00:44:03]Alessio: How do you decide what to give access to? So you have things like an expand prompt, which Dally 3 just does. It doesn't let you decide whether you should or not. [00:44:13]Swyx: As in, like, rewrites your prompts for you. [00:44:15]Suhail: Yeah, for that feature, I think once we get it to be cheaper, we'll probably just give it up. We'll probably just give it away. But we also decided something that might be a little bit different. We noticed that most of image generation is just, like, kind of casual. You know, it's in WhatsApp. It's, you know, it's in a Discord bot somewhere with Majorny. It's in ChatGPT. One of the differentiators I think we provide is at the expense of just lots of users necessarily. Mainstream consumers is that we provide as much, like, power and tweakability and configurability as possible. So the only reason why it's a toggle, because we know that users might want to use it and might not want to use it. There's some really powerful power user hobbyists that know what they're doing. And then there's a lot of people that just want something that looks cool, but they don't know how to prompt. And so I think a lot of Playground is more about going after that core user base that, like, knows, has a little bit more savviness and how to use these tools. You know, the average Dell user is probably not going to use ControlNet. They probably don't even know what that is. And so I think that, like, as the models get more powerful, as there's more tooling, hopefully you'll imagine a new sort of AI-first graphics editor that's just as, like, powerful and configurable as Photoshop. And you might have to master a new kind of tool. [00:45:28]Swyx: There's so many things I could go bounce off of. One, you mentioned about waiting. We have to kind of somewhat address the elephant in the room. Consistency models have been blowing up the past month. How do you think about integrating that? Obviously, there's a lot of other companies also trying to beat you to that space as well. [00:45:44]Suhail: I think we were the first company to integrate it. Ah, OK. [00:45:47]Swyx: Yeah. I didn't see your demo. [00:45:49]Suhail: Oops. Yeah, yeah. Well, we integrated it in a different way. OK. There are, like, 10 companies right now that have kind of tried to do, like, interactive editing, where you can, like, draw on the left side and then you get an image on the right side. We decided to kind of, like, wait and see whether there's, like, true utility on that. We have a different feature that's, like, unique in our product that is called preview rendering. And so you go to the product and you say, you know, we're like, what is the most common use case? The most common use case is you write a prompt and then you get an image. But what's the most annoying thing about that? The most annoying thing is, like, it feels like a slot machine, right? You're like, OK, I'm going to put it in and maybe I'll get something cool. So we did something that seemed a lot simpler, but a lot more relevant to how users already use these products, which is preview rendering. You toggle it on and it will show you a render of the image. And then graphics tools already have this. Like, if you use Cinema 4D or After Effects or something, it's called viewport rendering. And so we try to take something that exists in the real world that has familiarity and say, OK, you're going to get a rough sense of an early preview of this thing. And then when you're ready to generate, we're going to try to be as coherent about that image that you saw. That way, you're not spending so much time just like pulling down the slot machine lever. I think we were the first company to actually ship a quick LCM thing. Yeah, we were very excited about it. So we shipped it very quick. Yeah. [00:47:03]Swyx: Well, the demos I've been seeing, it's not like a preview necessarily. They're almost using it to animate their generations. Like, because you can kind of move shapes. [00:47:11]Suhail: Yeah, yeah, they're like doing it. They're animating it. But they're sort of showing, like, if I move a moon, you know, can I? [00:47:17]Swyx: I don't know. To me, it unlocks video in a way. [00:47:20]Suhail: Yeah. But the video models are already so much better than that. Yeah. [00:47:23]Swyx: There's another one, which I think is general ecosystem of Loras, right? Civit is obviously the most popular repository of Loras. How do you think about interacting with that ecosystem? [00:47:34]Suhail: The guy that did Lora, not the guy that invented Loras, but the person that brought Loras to Stable Diffusion actually works with us on some projects. His name is Simu. Shout out to Simu. And I think Loras are wonderful. Obviously, fine tuning all these Dreambooth models and such, it's just so heavy. And it's obvious in our conversation around styles and vibes, it's very hard to evaluate the artistry of these things. Loras give people this wonderful opportunity to create sub-genres of art. And I think they're amazing. Any graphics tool, any kind of thing that's expressing art has to provide some level of customization to its user base that goes beyond just typing Greg Rakowski in a prompt. We have to give more than that. It's not like users want to type these real artist names. It's that they don't know how else to get an image that looks interesting. They truly want originality and uniqueness. And I think Loras provide that. And they provide it in a very nice, scalable way. I hope that we find something even better than Loras in the long term, because there are still weaknesses to Loras, but I think they do a good job for now. Yeah. [00:48:39]Swyx: And so you would never compete with Civit? You would just kind of let people import? [00:48:43]Suhail: Civit's a site where all these things get kind of hosted by the community, right? And so, yeah, we'll often pull down some of the best things there. I think when we have a significantly better model, we will certainly build something that gets closer to that. Again, I go back to saying just I still think this is very nascent. Things are very underpowered, right? Loras are not easy to train. They're easy for an engineer. It sure would be nicer if I could just pick five or six reference images, right? And they might even be five or six different reference images that are not... They're just very different. They communicate a style, but they're actually like... It's like a mood board, right? And you have to be kind of an engineer almost to train these Loras or go to some site and be technically savvy, at least. It seems like it'd be much better if I could say, I love this style. Here are five images and you tell the model, like, this is what I want. And the model gives you something that's very aligned with what your style is, what you're talking about. And it's a style you couldn't even communicate, right? There's n

#9vor9 - Die Digitalthemen der Woche
127 - KI Tools, Teil 2: Bild

#9vor9 - Die Digitalthemen der Woche

Play Episode Listen Later Dec 19, 2023 26:21


Auch die letzte Episode im Jahr 2023 widmen wir DEM Tech Thema dieses Jahres: der Künstlichen Intelligenz. Nach der Textgenerierung in der Episode 126 geht es diesmal um die Generierung und auch die Manipulation von Bildern. Und auch diese Woche ist Lars der eher Skeptische und hofft, von Stefan etwas zu lernen, der sich in Sachen Bild-KI regelmäßig austobt. Hört selbst.

ChatGPT: News on Open AI, MidJourney, NVIDIA, Anthropic, Open Source LLMs, Machine Learning
Ideogram Debuts AI Image Generator with Stunning Custom Typography

ChatGPT: News on Open AI, MidJourney, NVIDIA, Anthropic, Open Source LLMs, Machine Learning

Play Episode Listen Later Oct 28, 2023 7:35


Tune in to explore the groundbreaking debut of Ideogram's AI Image Generator, an innovative tool that seamlessly combines art and custom typography for stunning visuals. Learn how this cutting-edge technology is transforming the creative landscape and making high-quality graphic design accessible to all. Discover the magic of AI-driven artistry and its impact on design in this must-listen episode. Get on the AI Box Waitlist: https://AIBox.ai/Join our ChatGPT Community: ⁠https://www.facebook.com/groups/739308654562189/⁠Follow me on Twitter: ⁠https://twitter.com/jaeden_ai⁠

WokeNFree
Episode 321: The Last Supper with AI

WokeNFree

Play Episode Listen Later Oct 25, 2023 49:51


This week's episode is all about AI-generated art. Share your thoughts in the comments below!  P.S. Have you signed up for Copy.ai yet?    Enjoy the series HERE Episode shout-out to MAKE USE OF, AI Robots, LeBron James in the Upside Down, WokeNFree Episode 242: Are You Using Copy.ai?, Bing Image Creator, Ideogram.ai, Canva AI generation, RoboCup 2023, and Colecteurs Download and use Newsly on www.newsly.me today!       Music Intro/Outro: “Thoughts” by Killah Smilez Music Outro: Wanderlust by noxz Make sure you check out the Killah Smilez song on Amazon Catch the music video by Killah Smilez HERE We're always working on new products and ideas, but sometimes it takes a little extra cash to bring them to life. Your financial support for the work we do means the world to us! Donate HERE!   ----more---- Shop WokeNFree Designs   Create your own Bonfire Shop Today! Get our book HERE Looking for a new video game to play? Check out these recommendations HERE Check out our course on the Law of Attraction HERE Get 10% off Saint Saxon Sound Swag with coupon code: WokeNFree10  Need advice? Connect with Natasha HERE Want to share the episode? Please share the episode on Facebook, Instagram, Twitter, YouTube, Pinterest, TikTok, and Soundcloud   Don't forget to subscribe to WokeNFree on iTunes, Stitcher, TuneIn, iHeartRadio, and Google Play Do you want to join the show as a guest on an upcoming episode? Contact us HERE Don't forget to submit a scenario to us for SCENARIO TIME!     Looking for cool new music to add to your content? Check out Uppbeat today!   Making content videos? GetMunch.com!  SCENARIO TIME: How would you respond to these scenarios in SCENARIO TIME? Let's chat HERE!  Have you reviewed our show yet? Pick your platform of choice HERE      Do you want to start a podcast? We are here to HELP! Schedule a FREE strategy session with us HERE This post contains affiliate links. That means if you click on a link and buy something, WokeNFree will earn a small commission from the advertiser at no additional cost to you.

Everyday AI Podcast – An AI and ChatGPT Podcast
EP 129: AI Image Generators - The Good, The Bad, and The Awesome

Everyday AI Podcast – An AI and ChatGPT Podcast

Play Episode Listen Later Oct 24, 2023 40:38


There are so many amazing and powerful AI image generators. From industry leaders like Midjourney and DALL-E to newer image generators getting released every week. So which one is right for you? Leonard Rodman, ChatGPT and Midjourney Consultant at Rodman.ai, joins us to go over the good, bad, and awesome image generators and how to prompt them to get what you're looking for.Newsletter: Sign up for our free daily newsletterMore on this Episode: Episode PageJoin the discussion: Ask Leonard and Jordan questions about AI image generatorsUpcoming Episodes: Check out the upcoming Everyday AI Livestream lineupWebsite: YourEverydayAI.comEmail The Show: info@youreverydayai.comConnect with Jordan on LinkedInTimestamps:[00:01:15] Daily AI news[00:03:40] About Leonard and Rodman.ai[00:08:35] Good beginner AI image generator[00:14:05] Ideogram - one of the originals[00:17:00] Recognizing AI-generated photos[00:25:00] Midjourney  breakdown[00:31:15] Audience questionsTopics Covered in This Episode:1. Introduction to AI Image Generators2. The Evolution of AI Image Generators3. Techniques and Tips for Effective AI Image Generation4. Identifying AI-Generated Images and Legal ConsiderationsKeywords:AI image generators, image generation, Leonardo, MidJourney, AI art, AI-generated images, AI technology, image recognition, digital photography, AI advancements, AI-generated storytelling, digital image manipulation, copyright laws, AI-generated venture capital pitches, marketing and advertising, image composition, AI-generated text, image modification, image quality, AI-generated photography, legal issues, DALL-E, digital cameras, AI-generated marketing, AI-generated storytelling, style subsets, image prompts, camera specifications, image ownership, image copyright, AI in different fields Get more out of ChatGPT by learning our PPP method in this live, interactive and free training! Sign up now: https://youreverydayai.com/ppp-registration/

AI Hustle: News on Open AI, ChatGPT, Midjourney, NVIDIA, Anthropic, Open Source LLMs
Ideogram Unveils AI Image Generator with Stunning Custom Typography

AI Hustle: News on Open AI, ChatGPT, Midjourney, NVIDIA, Anthropic, Open Source LLMs

Play Episode Listen Later Oct 4, 2023 7:45


Discover the creative marvels of Ideogram's new AI Image Generator in this episode. Join us as we delve into the world of custom typography, showcasing how this innovative tool can transform your visual content. Don't miss the insights and inspiration brought to you by Ideogram's cutting-edge technology. Get on the AI Box Waitlist: https://AIBox.ai/Join our ChatGPT Community: ⁠https://www.facebook.com/groups/739308654562189/⁠Follow me on Twitter: ⁠https://twitter.com/jaeden_ai⁠

SERP's Up SEO Podcast
SERP's Up+ | How marketers can best leverage AI (Bonus Episode)

SERP's Up SEO Podcast

Play Episode Listen Later Oct 2, 2023 37:11


What role does AI play in modern marketing? Can you leverage AI in your marketing? How do you use AI the ‘right' way?  Wix's Mordy Oberstien and Crystal Carter are joined by best-selling author and marketing influencer, Kim Garst, to discuss leveraging AI to expand your creativity and knowledge base in marketing. Kim discusses her experience using AI to fill certain skill gaps and shares her preferred chatbox for marketing purposes. Plus, learn all of the tools you need to be using to maximize your marketing efforts.  Artificial intelligence meets marketing in this episode of the SERP's Up+ Podcast!   Key Segments: [00:02:28] What's On This Episode of SERP's Up? [00:03:51] Focus Topic of the Week: The Role of AI in Marketing [00:04:26] Focus Topic Guest: Kim Garst   Hosts, Guests, & Featured People: Mordy Oberstein Crystal Carter Kim Garst   Resources: SERP's Up Podcast Wix SEO Learning Hub Searchlight SEO Newsletter Kim Garst Consulting Claude 2 vs GPT-4 in 2023: Comparing the Top AI Models ChatGPT  Claude AI Ideogram Midjourney Supermeme Dumme

Kudo's Radio -クドラジ-

Adobe Fireflyもある程度は文字出せるんだけど、Ideogramのクオリティには敵わないんだよな~

Outrage Factory
245: BB&B

Outrage Factory

Play Episode Listen Later Sep 22, 2023 52:56


What made the internet angry this Week?More Pumpkin Spice talk from your favorite basic bros. How gross was the Nick Chubb injury. Is Elon going to make people pay for twitter? Doubt it. Paul “Biznasty” Bissonette inadvertently cancels Babcock. What could people tell about our personalities by looking at our camera rolls. Horny Boebert is the Hallmark movie we all want to see. Who was the sexiest actress in Beetlejuice. Ideogram is the best AI program because you can make horny pics of famous people and use text. Russel Brand just got turbo canceled, how did he not get caught up in the OG MeToo movement. When someone gets canceled this fast is it previously set up and there's an emergency eject button? Lamentations of the modern Dad, those jerk boomers had it so easy in yet another way.  Find us:Go Fund Me https://gofund.me/d8eaa20fWeb outragefactory.comTwitter @OutrageFactPodInsta @outrage_factoryTik Tok @dalederuiterFacebook www.facebook.com/outragefactpodReddit r/OutragefactorypodEmail Outragefactpod@gmail.comCheck out our redbubble swag https://www.redbubble.com/shop/ap/49661850

Two Chairs Talking
Episode 96: The manner of your vile outrageous crimes

Two Chairs Talking

Play Episode Listen Later Sep 19, 2023 74:03


Perry and David talk about their latest reading, listening and watching in the crime genre, from a classic PI novel to the latest best-seller. Introduction (05:56) General News (11:06) Ned Kelly Awards 2023 (02:13) Dragon Awards 2023 (07:00) Queensland Literary Awards 2023 (01:43) What we've been reading and watching (55:24) In the Morning I'll Be Gone by Adrian McKinty (11:43) No Plan B by Lee Child and Andrew Child (11:46) The Sisters by Dervla McTiernan (05:01) D Is For Deadbeat by Sue Grafton (08:54) Down Cemetery Road by Mick Herron (07:58) The Lincoln Lawyer (Season 2) (06:01) Deadloch (03:54) Windup (01:10) Image generated by Ideogram.ai "Arrogant priest, this place demands I keep calm,Or thou shouldst find thou hast dishonor'd me.or you would be punished for your insult.Think not, although in writing I preferr'dDo not think that just because I wrote downThe manner of thy vile outrageous crimes,the details of your revolting terrible crimesThat therefore I have forg'd, or am not ablethat I have made anything up or cannotverbatim to rehearse the method of my pen." Henry VI Part I, Act 3 Scene 1

Two Chairs Talking
Episode 96: The manner of your vile outrageous crimes

Two Chairs Talking

Play Episode Listen Later Sep 19, 2023 74:03


Perry and David talk about their latest reading, listening and watching in the crime genre, from a classic PI novel to the latest best-seller. Introduction (05:56) General News (11:06) Ned Kelly Awards 2023 (02:13) Dragon Awards 2023 (07:00) Queensland Literary Awards 2023 (01:43) What we've been reading and watching (55:24) In the Morning I'll Be Gone by Adrian McKinty (11:43) No Plan B by Lee Child and Andrew Child (11:46) The Sisters by Dervla McTiernan (05:01) D Is For Deadbeat by Sue Grafton (08:54) Down Cemetery Road by Mick Herron (07:58) The Lincoln Lawyer (Season 2) (06:01) Deadloch (03:54) Windup (01:10) Click here for more info and indexes. Image generated by Ideogram.ai "Arrogant priest, this place demands I keep calm, Or thou shouldst find thou hast dishonor'd me. or you would be punished for your insult. Think not, although in writing I preferr'd Do not think that just because I wrote down The manner of thy vile outrageous crimes, the details of your revolting terrible crimes That therefore I have forg'd, or am not able that I have made anything up or cannot verbatim to rehearse the method of my pen." Henry VI Part I, Act 3 Scene 1

Let's Talk AI
#136 - Claude Pro, Ideogram, Chinese ChatGPT bots, Falcon 180B, RLAIF, export restrictions, Ghostwriter

Let's Talk AI

Play Episode Listen Later Sep 10, 2023 62:44


Our 136th episode with a summary and discussion of last week's big AI news! With guest host Daniel Bashir. Check out his AI interview podcast! Read out our text newsletter and comment on the podcast at https://lastweekin.ai/ Email us your questions and feedback at contact@lastweekin.ai Check out our sponsor, the SuperDataScience podcast. You can listen to SDS across all major podcasting platforms (e.g., Spotify, Apple Podcasts, Google Podcasts) plus there's a video version on YouTube. Timestamps + links: (00:00) Intro  (01:15) SuperDataScience Ad (01:51) Response to listeners Tools & Apps(02:47) Anthropic's Claude AI chatbot gets a paid plan for heavy users (04:36) Watch out, Midjourney! Ideogram launches AI image generator with impressive typography (06:37) Intuit launches generative AI–powered digital assistant for small businesses and consumers (07:17) Zoom Is Jumping on the AI Chatbot Bandwagon Applications & Business(08:47) China lets Baidu, others launch ChatGPT-like bots to public, tech shares jump (11:23) Tencent releases AI model for businesses as competition in China heats up (12:00) Microsoft says it will take the heat if Copilot AI users get sued (14:52) How We Chose the TIME100 Most Influential People in AI (17:30) ChatGPT creator OpenAI is reportedly earning $80M a month (19:16) AI chip startup d-Matrix raises $110 mln with backing from Microsoft (21:00) ThetaRay nabs $57M for AI tools to ID and fight money laundering (22:18) Sapeon raises $46m for AI chips (23:20) Imbue raises $200M to build AI models that can ‘robustly reason' Projects & Open Source(25:48) Announcing the commercial relicensing and expansion of DINOv2, plus the introduction of FACET (27:02) UAE launches Arabic large language model in Gulf push into generative AI (29:23) New Open-Source ‘Falcon' AI Language Model Overtakes Meta and Google Research & Advancements(31:01) Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities (33:57) RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback (36:22) Perception, performance, and detectability of conversational artificial intelligence across 32 university courses (39:28) SyncDreamer: Generating Multiview-consistent Images from a Single-view Image Policy & Safety(41:00) US curbs AI chip exports from Nvidia and AMD to some Middle East countries (42:55) China suspected of using AI on social media to sway US voters, Microsoft says (46:37) Trusting A.I.-written mushroom hunting guides sold on Amazon could get you killed. But like deadly fungi, identifying them is tricky (48:25) Ads for AI sex workers are flooding Instagram and TikTok (50:30) The UK releases key ambitions for global AI summit Synthetic Media & Art(51:55) AI Took the Stage at the World's Largest Arts Festival. Here's What Happened (54:12) Ghostwriter Returns With an A.I. Travis Scott Song, and Industry Allies (56:17) Artists sign open letter saying generative AI is good, actually (58:57) The latest canvas for Refik Anadol's AI-generated art? The new Sphere in Las Vegas (01:01:20) Outro

AI For Humans
Hot AI Art Tool Ideogram, Apple's Big Move & Google's 'Take a Meeting' AI | AI FOR HUMANS

AI For Humans

Play Episode Listen Later Sep 7, 2023 25:44


Today: Ideogram takes the AI world by storm, Apple's new AI push, Google's Meet will have your AI taking meetings for you, generative AI is ok in the Epic Games store & more! #artificialintelligence #ideogram #googleduet Hey there, AI For Human-ites! This week, we're navigating the buzzing world of AI, bringing you the latest updates and a sprinkle of real talk. Sadly, Kevin is taking a brief hiatus due to contracting COVID, but we're keeping the energy vibrant and the dialogue engaging. We're eagerly awaiting his return, but until then, we've curated a slightly shorter yet insightful episode for you!

Live On Tape Delay
Episode 439 - If You Want To Live...

Live On Tape Delay

Play Episode Listen Later Sep 4, 2023 57:22


Chris finally gets off the toilet, Rob smells amazing and John gets on a boat.  This is all before riviting conversations regarding water filters, SPEBSQSA and the new MAX docuseries "Telemarketers". And finally, to appease the AI overlords the boys make a few videos with the help of Invideo AI's AI co-pilot and create some album covers with Ideogram.   Enjoy!!

Front End Toolbox
Latest AI Tools: Ideogram, OpenAI's Mega Prompts, Loom AI, and More!

Front End Toolbox

Play Episode Listen Later Sep 4, 2023 12:23


John Siwicki is the host of Stack Snacks, a show that aims to make learning about AI simple and fun. He provides insights and commentary on various AI-related topics and shares interesting links and resources with his audience.Summary:In this episode of Stack Snacks, John Siwicki discusses several interesting AI-related topics. He starts by introducing Ideogram, an image-to-text generator that handles text in a unique way. He highlights some of the interesting results it produces, such as generating T-shirt designs and novelty items. John also mentions that Ideogram is free to try, but users should be cautious as their prompts are public-facing.Next, John talks about OpenAI's guide for teachers on how to use GPT in the classroom. He emphasizes that even though it's targeted at teachers, it provides valuable insights and thought exercises for everyone. The guide covers topics like how GPT works, its limitations, and biases. John encourages his audience to check it out and explore different ways of writing prompts.Moving on, John discusses Loom AI, a new feature added to the screen recording app Loom. He highlights the various enhancements Loom AI offers, such as automatically generating titles, creating summarizations, adding chapters, and providing editing features similar to Descript. John finds the addition of AI-powered talking points and the ability to generate personalized versions of recordings with variables particularly interesting. He mentions that Loom AI is available as an add-on with a monthly fee.John then introduces Audio Read, a tool that converts articles into voiceovers. He mentions that what sets Audio Read apart is its ability to generate an RSS feed, allowing users to subscribe to the voiceovers in their podcast player. He sees this as a convenient way to consume articles while on the go.Lastly, John briefly mentions a new Canva plugin called Chat GPT. Although he couldn't find much information about it, he demonstrates how it integrates with Canva and allows users to create logos based on prompts.Key Takeaways:* Ideogram is a text-to-image generator that produces high-quality and interesting results.* OpenAI's guide for teachers on using GPT in the classroom provides valuable insights for everyone, not just teachers.* Loom AI offers several enhancements to the screen recording app, including automatic title generation, summarizations, chapters, and advanced editing features.* Audio Read converts articles into voiceovers and provides an RSS feed for easy podcast-style consumption.* Canva has a new Chat GPT plugin that integrates with the design platform and allows users to create logos based on prompts.Quotes:* "Ideogram is a high-quality, free-to-use text-to-image generator that can produce interesting results."* "OpenAI's guide for teachers on using GPT in the classroom provides valuable insights and thought exercises for everyone."* "Loom AI offers enhancements like automatic title generation, summarizations, chapters, and advanced editing features."* "Audio Read converts articles into voiceovers and provides an RSS feed for easy podcast-style consumption."* "Canva's Chat GPT plugin allows users to create logos based on prompts and integrates seamlessly with the design platform." This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.stack-snacks.com

KI-Update – ein Heise-Podcast
KI-Update kompakt: ChatGPT, Förderprogramm, Ideogram, Medizin

KI-Update – ein Heise-Podcast

Play Episode Listen Later Aug 31, 2023 11:22


ChatGPT Copyright-Klage Förderprogramm für Open-Source-KI-Projekten Kostenlose Midjourney-Alternative Ideogram Und mit KI wieder Sprechen lernen heise.de/ki-update https://www.heise.de/thema/Kuenstliche-Intelligenz https://the-decoder.de/ https://www.heiseplus.de/podcast

No me da la vida
2.5 - Para hablar de humor tech con Rita Iglesias

No me da la vida

Play Episode Listen Later Aug 31, 2023 91:57


En el episodio 2.5 os hablamos de humor tech, de nuevos usos de IA como Cloudinary AI, OverflowAI, la IA de Google Labs en Gmail, de pólemicas con Zoom, Brave, Elon Musk eliminando la opción de bloquear, el NO hackeo de Discord, de Python en Excel, del rediseño de Slack, del nuevo CodeSandbox por Google: Project IDX, de Anchor Positioning, de Jam.dev, de Ideogram, entre otras muchas cosas

Kudo's Radio -クドラジ-

他言語対応したら神サービスになりますね

AI News Briefing
AI Killer Drones, Germany's AI Funding and Ideogram AI | AI News Briefing | 28.08.2023

AI News Briefing

Play Episode Listen Later Aug 28, 2023


This is the AI News Briefing of August 28, 2023.(00:29) Germany doubles AI research funding(01:16) Ideogram AI vs. Midjourney in text-to-image(02:00) U.S. Air Force's AI-powered Valkyrie DronesGermany doubles AI research funding: https://www.reuters.com/technology/germany-plans-double-ai-funding-race-with-china-us-2023-08-23/Ideogram AI vs. Midjourney in text-to-image: https://ideogram.ai/U.S. Air Force's AI-powered Valkyrie Drones: https://www.reuters.com/technology/germany-plans-double-ai-funding-race-with-china-us-2023-08-23/Follow our newsletter at www.adepto.ai for a deeper dive into these fascinating developments and for the latest AI news and insights.The AI News Briefing has been produced by Adepto in cooperation with Wondercraft AI.Music: Inspire by Kevin MacLeod (incompetech.com), Licensed under Creative Commons: By Attribution 3.0 http://creativecommons.org/licenses/by/3.0/

AI Chat: ChatGPT & AI News, Artificial Intelligence, OpenAI, Machine Learning
Ideogram Launches AI Image Generator With (Good) Custom Typography

AI Chat: ChatGPT & AI News, Artificial Intelligence, OpenAI, Machine Learning

Play Episode Listen Later Aug 27, 2023 7:52


In this episode, we dive into Ideogram's latest launch—an AI-driven image generator that's turning heads for its innovative approach to typography. We explore how this game-changing technology is redefining graphic design. Get on the AI Box Waitlist: ⁠https://AIBox.ai/⁠ Facebook Community: ⁠⁠https://www.facebook.com/groups/739308654562189/⁠⁠ Discord Community: ⁠https://aibox.ai/discord⁠ Follow me on X: ⁠https://twitter.com/jaeden_ai⁠