Podcasts about generativeai

  • 186PODCASTS
  • 535EPISODES
  • 32mAVG DURATION
  • 5WEEKLY NEW EPISODES
  • Sep 19, 2025LATEST

POPULARITY

20172018201920202021202220232024


Best podcasts about generativeai

Latest podcast episodes about generativeai

AWS for Software Companies Podcast
Ep147: Securing Generative AI Investigations Against Invisible Risks & Threats w Cohesity

AWS for Software Companies Podcast

Play Episode Listen Later Sep 19, 2025 21:34


Aditya Vasudevan, Cohesity's cyber recovery expert, shares battle-tested insights from defending Fortune 100 companies against AI-powered cyberattacks.Topics Include:Cohesity protects 85% of Fortune 100 data with battle-tested cyber recovery experienceTop 10 cyber adversaries target organizations; Cohesity has defended against most major threatsGenAI adopted by 100 million users in two months, creating unprecedented security challengesNew AI threats include prompt injection, synthetic identities, shadow AI, and supply vulnerabilitiesAttackers now use AI for sophisticated phishing, automated malware, and accelerated attack chainsReal companies completely banned AI after code leaks, misuse incidents, and data concernsThree-pillar security approach: fight AI with AI, enhanced training, and automated workflowsSecure AI design requires private deployments, complete traceability, and role-based access controlsAmazon Bedrock offers built-in guardrails, private VPCs, and enterprise monitoring capabilitiesCohesity's Gaia demonstrates secure AI with RAG architecture and permission-aware data accessResilience strategy combines immutable backups, anomaly detection, and recovery automation for incidentsProper AI security reduces cyber insurance premiums and prevents costly downtime disastersParticipants:Aditya Vasudevan - GVP of Cyber Resiliency, Cohesity Further Links:Cohesity: Website | LinkedIn | AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep146: Strategies to enhance organizational security culture with Sonatype

AWS for Software Companies Podcast

Play Episode Listen Later Sep 17, 2025 15:22


Tyler Warden, SVP of Product at Sonatype, shares surprising research on security, productivity and prioritization, with actionable strategies for organizational transformation. Topics Include:Tyler from Sonatype (Maven creators) shares research on security culture in developmentSecurity is more cultural than tooling, with rising supply chain attacksDevelopment speeds up while global regulations rapidly change across marketsTyler's background: wanted to be a Broadway conductor, not tech speakerBeethoven's 9th Symphony story: nephew missed a dot, changing tempo foreverWe can "be the dot" - small changes creating big organizational impactThree organization types: Leaders (collaborative), Adapters (balanced), Protectors (security-first)Leaders achieve best productivity and security but face executive skepticismResearch reveals balanced teams outperform purely security-focused or productivity-focused approachesHigh-performance teams go faster AND stay more secure than alternatives"Yes" philosophy from improv comedy: fun happens when we enable innovationApply proven supply chain principles from manufacturing to software development security Participants:Tyler Warden – Senior Vice President, Product, SonatypeFurther Links:Sonatype: Website | LinkedIn | AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Oracle University Podcast
Oracle's AI Ecosystem

Oracle University Podcast

Play Episode Listen Later Sep 16, 2025 15:39


In this episode, Lois Houston and Nikita Abraham are joined by Principal Instructor Yunus Mohammed to explore Oracle's approach to enterprise AI. The conversation covers the essential components of the Oracle AI stack and how each part, from the foundational infrastructure to business-specific applications, can be leveraged to support AI-driven initiatives.   They also delve into Oracle's suite of AI services, including generative AI, language processing, and image recognition.     AI for You: https://mylearn.oracle.com/ou/course/ai-for-you/152601/   Oracle University Learning Community: https://education.oracle.com/ou-community   LinkedIn: https://www.linkedin.com/showcase/oracle-university/   X: https://x.com/Oracle_Edu   Special thanks to Arijit Ghosh, David Wright, Kris-Ann Nansen, Radhika Banka, and the OU Studio Team for helping us create this episode.   -------------------------------------------------------------   Episode Transcript:  00:00 Welcome to the Oracle University Podcast, the first stop on your cloud journey. During this series of informative podcasts, we'll bring you foundational training on the most popular Oracle technologies. Let's get started! 00:25 Lois: Welcome to the Oracle University Podcast! I'm Lois Houston, Director of Innovation Programs with Oracle University, and with me is Nikita Abraham, Team Lead: Editorial Services. Nikita: Hey everyone! In our last episode, we discussed why the decision to buy or build matters in the world of AI deployment. Lois: That's right, Niki. Today is all about the Oracle AI stack and how it empowers not just developers and data scientists, but everyday business users as well. Then we'll spend some time exploring Oracle AI services in detail.  01:00 Nikita: Yunus Mohammed, our Principal Instructor, is back with us today. Hi Yunus! Can you talk about the different layers in Oracle's end-to-end AI approach? Yunus: The first base layer is the foundation of AI infrastructure, the powerful compute and storage layer that enables scalable model training and inferences. Sitting above the infrastructure, we have got the data platform. This is where data is stored, cleaned, and managed. Without a reliable data foundation, AI simply can't perform. So base of AI is the data, and the reliable data gives more support to the AI to perform its job. Then, we have AI and ML services. These provide ready-to-use tools for building, training, and deploying custom machine learning models. Next, to the AI/ML services, we have got generative AI services. This is where Oracle enables advanced language models and agentic AI tools that can generate content, summarize documents, or assist users through chat interfaces. Then, we have the top layer, which is called as the applications, things like Fusion applications or industry specific solutions where AI is embedded directly into business workflows for recommendations, forecasting or customer support. Finally, Oracle integrates with a growing ecosystem of AI partners, allowing organizations to extend and enhance their AI capabilities even further. In short, Oracle doesn't just offer AI as a feature. It delivers it as a full stack capability from infrastructure to the layer of applications. 02:59 Nikita: Ok, I want to get into the core AI services offered by Oracle Cloud Infrastructure. But before we get into the finer details, broadly speaking, how do these services help businesses? Yunus: These services make AI accessible, secure, and scalable, enabling businesses to embed intelligence into workflows, improve efficiency, and reduce human effort in repetitive or data-heavy tasks. And the best part is, Oracle makes it easy to consume these through application interfaces, APIs, software development kits like SDKs, and integration with Fusion Applications. So, you can add AI where it matters without needing a data scientist team to do that work.  03:52 Lois: So, let's get down to it. The first core service is Oracle's Generative AI service. What can you tell us about it?  Yunus: This is a fully managed service that allows businesses to tap into the power of large language models. You can actually work with these models from scratch to a well-defined develop model. You can use these models for a wide range of use cases like summarizing text, generating content, answering questions, or building AI-powered chat interfaces.  04:27 Lois: So, what will I find on the OCI Generative AI Console? Yunus: OCI Generative AI Console highlights three key components. The first one is the dedicated AI cluster. These are GPU powered environments used to fine tune and host your own custom models. It gives you control and performance at scale. Then, the second point is the custom models. You can take a base language model and fine tune it using your own data, for example, company manuals or HR policies or customer interactions, which are your own personal data. You can use this to create a model that speaks your business language. And last but not the least, the endpoints. These are the interfaces through which your application connect to the model. Once deployed, your app can query the model securely and at different scales, and you don't need to be a developer to get started. Oracle offers a playground, which is a non-core environment where you can try out models, craft parameters, and test responses interactively. So overall, the generative AI service is designed to make enterprise-grade AI accessible and customizable. So, fitting directly into business processes, whether you are building a smart assistant or you're automating the content generation process.  06:00 Lois: The next key service is OCI Generative AI Agents. Can you tell us more about it?  Yunus: OCI Generative AI agents combines a natural language interface with generative AI models and enterprise data stores to answer questions and take actions. The agent remembers the context, uses previous interactions, and retrieves deeper product speech details. They aren't just static chat bots. They are context aware, grounded in business data, and able to handle multi-turns, follow-up queries with relevant accurate responses, and driving productivity and decision-making across departments like sales, support, or operations. 06:54 Oracle University's Race to Certification 2025 is your ticket to free training and certification in today's hottest tech. Whether you're starting with Artificial Intelligence, Oracle Cloud Infrastructure, Multicloud, or Oracle Data Platform, this challenge covers it all! Learn more about your chance to win prizes and see your name on the Leaderboard by visiting education.oracle.com/race-to-certification-2025. That's education.oracle.com/race-to-certification-2025. 07:37 Nikita: Welcome back! Yunus, let's move on to the OCI Language service.  Yunus: OCI Language helps business understand and process natural language at scale. It uses pretrained models, which means they are already trained on large industry data sets and are ready to be used right away without requiring AI expertise. It detects over 100 languages, including English, Japanese, Spanish, and more. This is great for global business that receive multilingual inputs from customers. It works with identity sentiments. For different aspects of the sentence, for example, in a review like, “The food was great, but the service sucked,” OCI Language can tell that food has a positive sentiment while service has a negative one. This is called aspect-based sentiment analysis, and it is more insightful than just labeling the entire text as positive or negative. Then we have got to identify key phrases representing important ideas or subjects. So, it helps in extracting these key phrases, words, or terms that capture the core messages. They help automate tagging, summarizing, or even routing of content like support tickets or emails.  In real life, the businesses are using this for customer feedback analysis, support ticket routing, social media monitoring, and even regulatory compliances.  09:21 Nikita: That's fantastic. And what about the OCI Speech service?  Yunus: The OCI Speech is an AI service that transcribes speech to text. Think of it as an AI-powered transcription engine that listens to the spoken English, whether in audio or video files, and turns it into usable and searchable and readable text. It provides timestamps, so you know exactly when something was said. A valuable feature for reviewing legal discussions, media footages, or compliance audits. OCI Speech even understands different speakers. You don't need to train this from scratch. It is pre-trained model hosted on an API. Just send your audio to the service, and you get an accurate timestamp text back in return. 10:17 Lois: I know we also have a service for object detection… called OCI Vision?  Yunus: OCI Vision uses pretrained, deep learning models to understand and analyze visual content. Just like a human might, you can upload an image or videos, and the AI can tell you what is in it and where they might be useful. There are two primary use cases, which you can use this particular OCI Vision for. One is for object detection. You have got a red color car. So OCI Vision is not just identifying that's a car. It is detecting and labeling parts of the car too, like the bumper, the wheels, the design components. This is a critical in industries like manufacturing, retail, or logistics. For example, in quality control, OCI Vision can scan product images to detect missing or defective parts automatically.  Then we have got the image classification. This is useful in scenarios like automated tagging of photos, managing digital assets, classifying this particular scene or context of this particular scene. So basically, when we talk about OCI Vision, which is actually a fully managed, no complex model training is required for this particular service. It's available via API. It is also working with defining their own custom model for working with the environments. 11:51 Nikita: And the final service is related to text and called OCI Document Understanding, right? Yunus: So OCI Document Understanding allows businesses to automatically extract structured insights from unstructured documents like invoices, contracts, recipes, and also sometimes resumes, or even business documents. 12:13 Nikita: And how does it work? Yunus: OCI reads the content from the scanned document. The OCR is smarter. It recognizes both printed and handwritten text. Then determines what type of document it is. So document classification is done. Text recognition recognizes text, then classifies the document. For example, if this is a purchase order, or bank statement, or any medical report. If your business handles documents in multiple languages, then the AI can actually help in language detection also, which helps you in routing the language or translating that particular language. Many documents contain structured data in table format. Think pricing tables or line items. OCI will help you in extracting these with high accuracy for reporting on feeding into ERP systems. And finally, I would say the key value extraction. It puts our critical business values like invoice numbers, payment amounts, or customer names from fields that may not always allow a fixed format. So, this service reduces the need for manual review, cuts down processes time, and ensures high accuracy for your system. 13:36 Lois: What are the key takeaways our listeners should walk away with after this episode? Yunus: The first one, Oracle doesn't treat AI as just a standalone tool. Instead, AI is integrated from the ground up. Whether you're talking about infrastructure, data platforms, machine learning services, or applications like HCM, ERP, or CX. In real world, the Oracle AI Services prioritize data management, security, and governance, all essential for enterprise AI use cases. So, it is about trust. Can your AI handle sensitive data? Can it comply with regulations? Oracle builds its AI services with strong foundation in data governance, robust security measures, and tight control over data residency and access. So this makes Oracle AI especially well-suited for industries like health care, finance, logistics, and government, where compliance and control aren't optional. They are critical.   14:44 Nikita: Thank you for another great conversation, Yunus. If you're interested in learning more about the topics we discussed today, head on over to mylearn.oracle.com and search for the AI for You course.  Lois: In our next episode, we'll get into Predictive AI, Generative AI, Agentic AI, all with respect to Oracle Fusion Applications. Until then, this is Lois Houston… Nikita: And Nikita Abraham, signing off! 15:10 That's all for this episode of the Oracle University Podcast. If you enjoyed listening, please click Subscribe to get all the latest episodes. We'd also love it if you would take a moment to rate and review us on your podcast app. See you again on the next episode of the Oracle University Podcast.  

AWS for Software Companies Podcast
Ep145: Transforming the CFO's Office with Amazon Bedrock

AWS for Software Companies Podcast

Play Episode Listen Later Sep 15, 2025 31:51


Learn how Amazon Bedrock enabled Prophix to build enterprise-grade AI agents that transform secure CFO workflows, while delivering real-time financial intelligence through advanced agentic architecture. Topics Include:AWS and Prophix leaders discuss building autonomous AI agents for financial managementAI agents defined: autonomous systems that reason, plan, and execute tasks independentlyEvolution from simple chatbots to collaborative multi-agent systems solving complex problemsCore agent components: cognitive planning module, memory systems, and external tool integrationsProphix: 30-year financial software company serving 3,500 CFO offices globally across industriesProphix One Intelligence: platform-level AI service powering predictions, analysis, and automationCustomer concerns addressed: data privacy, role-based security, accuracy, and cost controlRejected "models in the sky" approach for AWS Bedrock's managed, controllable infrastructureAgentic architecture: LLMs generate API parameters instead of processing massive datasetsReal-time data access, automatic security inheritance, and A/B testing capabilities achievedLive demo: automated budgeting workflows, natural language queries, and autonomous task executionAWS introduces AgentCore platform to simplify agent development for enterprise customersParticipants:Anurag Yagnik – Chief Technology Officer, ProphixDeborshi Choudhury – Sr Solutions Architect – ISV, Amazon Web ServicesFurther Links:Prophix: Website | LinkedIn See how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep144: 8 Trillion Observations a Week: How Arctic Wolf Uses AI to Stop Ransomware Attacks

AWS for Software Companies Podcast

Play Episode Listen Later Sep 11, 2025 19:27


Dean Teffer of Arctic Wolf reveals how they process 8 trillion weekly security observations to find "a needle in a stack of needles," and breaks down real-world GenAI lessons learned.Topics Include:Dean Teffer, VP of AI at Arctic Wolf, discusses company's GenAI journeyArctic Wolf: decade-old security operations company serving mid-market customers globallyOperates massive security operation center, now launching AI-powered productsAI agent recently identified Black Basta ransomware attack, enabling rapid containmentDean's 15+ years in cybersecurity: traditional ML focused on detectionGenAI breakthrough allows natural language interaction with security modelsArctic Wolf processes 8 trillion weekly observations, correlating suspicious activitiesChallenge: finding specific threats in "stack of needles," not haystackSuccess measured by making human analysts faster, more consistent, scalableEvolved from treating GenAI like traditional ML to integrated workflowsKey misconception: GenAI isn't magic, needs proper data and reasoningAdvice: start with existing challenges, build flexible systems for adaptationGenAI excels at summarizing information and supporting complex decisionsFuture vision: AI handles routine threats, humans focus on creativityDemocratizing machine learning capabilities to broader range of subject expertsParticipants:Dean Teffer – Vice President of AI, Arctic WolfFurther Links:Arctic Wolf: Website | LinkedIn | AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep143: Beyond Passwords: CyberArk's Vision for Human, Machine, and AI Identity Security

AWS for Software Companies Podcast

Play Episode Listen Later Sep 10, 2025 21:32


CyberArk's technology leader discusses their strategy for securing against AI threats, protecting agentic AI systems, and their vision for the future in an increasingly AI-driven cybersecurity landscape.Topics Include:CyberArk celebrates recent exciting news while discussing their incredible cybersecurity journeyFounded in 1999, CyberArk pioneered privilege access management and expanded into comprehensive identity securityCompany executed textbook SaaS transformation from perpetual licensing to subscription-based cloud modelLeadership set clear customer expectations, framing SaaS shift as faster innovation deliveryAddressed customer concerns about cost predictability, security compliance, and data residency requirementsTechnical team implemented lift-and-shift architecture with AWS RDS and multi-tenant improvementsCorporate initiative tracked weekly metrics and milestones throughout full development lifecycle processCustomer Success evolved from transactional support to strategic partnership embedded in security journeysAWS partnership fundamental to cloud journey with 25+ integrations and Marketplace collaborationAI strategy focuses on three pillars: using AI, securing against AI threatsFuture 12-24 months: continue securing all identities while expanding AI capabilities and solutionsAWS partnership expanding in 2025 leveraging machine identity leadership and GenAI advancesParticipants:Peretz Regev – Chief Product & Technology Officer, CyberArkBoaz Ziniman – Principal Developer Advocate - EMEA, Amazon Web ServicesFurther Links:· CyberArk: Website – LinkedIn – AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Oracle University Podcast
Buy or Build AI?

Oracle University Podcast

Play Episode Listen Later Sep 9, 2025 15:58


How do you decide whether to buy a ready-made AI solution or build one from the ground up? The choice is more than just a technical decision; it's about aligning AI with your business goals.   In this episode, Lois Houston and Nikita Abraham are joined by Principal Instructor Yunus Mohammed to examine the critical factors influencing the buy vs. build debate. They explore real-world examples where businesses must weigh speed, customization, and long-term strategy. From a startup using a SaaS chatbot to a bank developing a custom fraud detection model, Yunus provides practical insights on when to choose one approach over the other.   AI for You: https://mylearn.oracle.com/ou/course/ai-for-you/152601/   Oracle University Learning Community: https://education.oracle.com/ou-community   LinkedIn: https://www.linkedin.com/showcase/oracle-university/   X: https://x.com/Oracle_Edu   Special thanks to Arijit Ghosh, David Wright, Kris-Ann Nansen, Radhika Banka, and the OU Studio Team for helping us create this episode.   ---------------------------------------------------------------   Episode Transcript: 00:00 Welcome to the Oracle University Podcast, the first stop on your cloud journey. During this series of informative podcasts, we'll bring you foundational training on the most popular Oracle technologies. Let's get started! 00:26 Nikita: Welcome to the Oracle University Podcast! I'm Nikita Abraham, Team Lead: Editorial Services with Oracle University, and with me is Lois Houston, Director of Innovation Programs. Lois: Hi there! Last week, we spoke about the key stages in a typical AI workflow and how data quality, feedback loops, and business goals influence AI success. 00:50 Nikita: In today's episode, we're going to explore whether you should buy or build AI apps. Joining us again is Principal Instructor Yunus Mohammed. Hi Yunus, let's jump right in. Why does the decision of buy versus build matter? Yunus: So when we talk about buy versus build matters, we need to consider the strategic business decisions over here. They are related to the strategic decisions which the business makes, and it is evaluated in the decision lens. So the center of the decision lens is the business objective, which identifies what are we trying to solve. Then evaluate our constraints based on that particular business objective like the cost, the time, and the talent. And finally, we can decide whether we need to buy or build. But remember, there is no single correct answer. What's right for one business may not be working for the other one. 01:54 Lois: OK, can you give us examples of both approaches? Yunus: The first example where we have got a startup using a SaaS AI chatbot. Now, being a startup, they have to choose a ready-made solution, which is an AI chatbot. Now, the question is, why did they do this? Because speed and simplicity mattered more than deep customization that is required for the chatbot. So, their main aim was to have it ready in short period of time and make it more simpler. And this actually lead them to get to the market fast with low upfront cost and minimal technical complexities. But in some situations, it might be different. Like, your bank, which needs to build a fraud model. It cannot be outsourced or got from the shelf. So, they build a custom model in-house. With this custom model, they actually have a tighter control, and it is tuned to their standards. And it is created by their experts. So these two generic examples, the chatbot and the fraud model example, helps you in identifying whether I should go for a SaaS product with simple choice of selecting an existing LLM endpoint and not making any changes. Or should I go with model depending on my business and organization requirement and fine tuning that model later to define a better implementation of the scenarios or conditions that I want to do which are specific to my organization. So here you decide with the reference whether I want it to be done faster, or whether I want to be more customized to my organization. So buy it, when it is generic, or build when it is strategic. The SaaS, which is basically software as a service, refers to ready to use cloud-based applications that you access via internet. You can log into the platform and use the built-in AI, there's no setup requirement for those. Real-world examples can be Oracle Fusion apps with AI features enabled. So in-house integration means embedding AI with my own requirements into your own systems, often using custom APIs, data pipelines, and hosting it. It gives you more flexibility but requires a lot of resources and expertise. So real-world example for this scenario can be a logistics heavy company, which is integrating a customer support model into their CX. 04:41 Lois: But what are the pros and cons of each approach? Yunus: So, SaaS and Fusion Applications, basically, they offer fast deployment with little to no coding required, making them ideal for business looking to get started quickly and faster. And they typically come with lower upfront costs and are maintained by vendor, which means updates, security, support are handled externally. However, there are limited customizations and are best suited for common, repeatable use cases. Like, it can be a standard chatbot, or it can be reporting tools, or off the shelf analytics that you want to use. But the in-house or custom integration, you have more control, but it takes longer to build and requires a higher initial investment. The in-house or custom integration approach allows full customization of the features and the workflows, enabling you to design and tailor the AI system to your specific needs. 05:47 Nikita: If you're weighing the choice between buying or building, what are the critical business considerations you'd need to take into account? Yunus: So let's take one of the key business consideration which is time to market. If your goal is to launch fast, maybe you're a startup trying to gain traction quickly, then a prebuilt plug and play AI solution, for example, a chatbot or any other standard analytical tool, might be your best bet. But if you have time and you are aiming for precision, a custom model could be worth the wait. Prebuilt SaaS tools usually have lower upfront costs and a subscription model. It works with putting subscriptions. Custom solutions, on the other hand, may require a bigger investment upfront. In development, you require high talent and infrastructures, but could offer cost savings in the long run. So, ask yourself a question here. Is this AI helping us stand out in the market? If the answer is yes, you may want to build something which is your proprietary. For example, an organization would use a generic recommendation engine. It's a part of their secret sauce. Some use cases require flexibility, like you want to tailor the rules to match your specific risk criteria. So, under that scenarios, you will go for customizing. So, you will go with off the shelf solutions may not give you deep enough requirements that you want to evaluate. So, you get those and you try to customize those. You can go for customization of your AI features. The other important key business consideration is the talent and expertise that your organization have. So, the question that you need to ask in the organization is, do you have an internal team who is well versed in developing AI solutions for you? Or do you have access to one of the teams which can help you build your own proprietary products? If not, you'll go with SaaS. If you do have, then building could unlock greater control over your AI features and AI models. The next core component is your security and data privacy. If you're handling sensitive information, like for example, the health care or finance data, you might not want to send your data to the third-party tools. So in-house models offer better control over data security and compliance. When we leverage a model, it could be a prebuilt or custom model. 08:50 Oracle University is proud to announce three brand new courses that will help your teams unlock the power of Redwood—the next generation design system. Redwood enhances the user experience, boosts efficiency, and ensures consistency across Oracle Fusion Cloud Applications. Whether you're a functional lead, configuration consultant, administrator, developer, or IT support analyst, these courses will introduce you to the Redwood philosophy and its business impact. They'll also teach you how to use Visual Builder Studio to personalize and extend your Fusion environment. Get started today by visiting mylearn.oracle.com.  09:31 Nikita: Welcome back! So, getting back to what you were saying before the break, what are pre-built and custom models? Yunus: A prebuilt model is an AI solution that has already been trained by someone else, typically a tech provider. It can be used to perform a specific task like recognizing images, translating text, or detecting sentiments. You can think of it like buying a preassembled appliance. You plug it in, configure a few settings, and it's ready to use. You don't need to know how the internal parts work. You benefit from the speed, ease, and reliability of this particular model, which is a prebuilt model. But you can't easily change how it works under the hood. Whereas, a custom model is an AI solution that your organization designs and trains and tunes specifically for their business problems using their own data. You can think of it like designing your own suit. It takes more time and effort to create. It is built to your exact measurements and needs. And you have full control over how it performs and evolves. 10:53 Lois: So, when would you choose a pre-built versus a custom model? Yunus: Depending on speed, simplicity, control, and customization, you can decide on using a prebuilt or to create a custom model. Prebuilt models are like plug and play solutions. Think of tools like Google Translate for languages. OpenAI APIs for summarizing sentiment analysis or chatbots, they are quick to deploy, require low technical effort, great for getting started fast, but they also have limits. Customization is very minimal, and you may not be able to fine tune it to your specific tone or business logic. These work well when the problem is common and nonstrategic, like, scanning documents or auto tagging images. The custom-build model, on the other hand, is a model that is built from the ground up. Using your own data and objectives, they take longer, and they require technical expertise. But they offer precise control, full alignment with your business needs. And these are ideal when you are dealing with sensitive data, competitive workflows, highly specific customer interactions. For example, a bank may build a custom model which can be used for fraud detection, which can be tuned to their exact transaction standards and the patterns of their transactions. 12:37 Nikita: What if someone wants the best of both worlds?  Yunus: We've also got a hybrid approach. In hybrid approach, we actually talk about the adaptation of AI with a strategy which is termed as hybrid strategy. Many companies today don't start by building AI from scratch. Instead, they begin with prebuilt models, like using an API, which can be already performing tasks like summarizing, translating, or answering questions using generic knowledge. This set will help you in getting up and running quickly with a small level results. As your business matures, you can start to layer in your custom data. Think internal policies, frequently asked questions, or customer interactions. And then you can fine tune the model to behave the way your business needs it to behave. Now, your AI starts producing business-ready output, smarter, more relevant, and aligned with your tone, brand, or compliance needs.  13:45 Lois: Ok…let's think of AI deployment in the hybrid approach as following a pyramid or ladder like structure. Can you take us through the different levels?  Yunus: So, on the top, quick start, minimal setup, great for business automation, which can be used as a pilot use case. So, if I'm taking off the shelf APIs or platforms, they can be giving me a faster, less set of requirements, and they are basically acting like a pilot use. Later, you can add your own data or logic so you can add your data. You can fine tune or change your business logic. And this is where fine tuning and prompt engineering helps tailor the AI to your workflows and your language. And then at the end, which is at the bottom, you build your own model. It is reserved for core capabilities or competitive advantages where total control and differentiation matters in building that particular model. You don't need to go all in from one day. So, start with what is available, like, use an off shelf, API, or platform, customize as you grow. Build only when it gives you a true edge. This is what we call the best of both worlds, build and buy. 15:05 Lois: Thank you so much, Yunus, for joining us again. To learn more about the topics covered today, visit mylearn.oracle.com and search for the AI for You course. Nikita: Join us next week for another episode of the Oracle University Podcast where we discuss the Oracle AI stack and Oracle AI services. Until then, this is Nikita Abraham… Lois: And Lois Houston, signing off! 15:29 That's all for this episode of the Oracle University Podcast. If you enjoyed listening, please click Subscribe to get all the latest episodes. We'd also love it if you would take a moment to rate and review us on your podcast app. See you again on the next episode of the Oracle University Podcast.

AWS for Software Companies Podcast
Ep142: Transforming ISV Businesses Through Modern Data Platforms with Coveo, DTEX Systems and Honeycomb

AWS for Software Companies Podcast

Play Episode Listen Later Sep 8, 2025 44:12


Three leading ISV executives from Coveo, DTEX Systems and Honeycomb, reveal how companies with proprietary datasets are gaining unbeatable competitive advantages in the AI era and share real-world strategies how you have similar outcomes.Topics Include:Panel introduces three ISV leaders discussing data platform transformation for AIDTEX focuses on insider threats, Coveo on enterprise search, Honeycomb on observabilityCompanies with proprietary datasets gain strongest competitive advantage in AI transformationData gravity concept: LLMs learning from unique datasets create defensible business positionsCoveo maintains unified enterprise index with real-time content and access rights syncHoneycomb enables subsecond queries for analyzing logs, traces, and metrics at scaleMulti-tenant architectures balance shared infrastructure benefits with single-tenant data separationCoveo deployed 140,000 times last year using mostly multi-tenant, some single-tenant componentsDTEX scaled from thousands to hundreds of thousands endpoints after architectural transformationCapital One partnership taught DTEX how to break monolithic architecture into servicesApache Iceberg and open table formats enable interoperability without data duplicationHoneycomb built custom format following similar patterns with hot/cold storage tiersBusiness data catalogs become critical for AI agents understanding dataset contextMCP servers allow AI systems to leverage structured cybersecurity datasets effectivelyDTEX used Cursor with their data to identify North Korean threat actorsReal-time AI data needs balanced with costs using right models for jobsCaching strategies and precise context reduce expensive LLM inference calls unnecessarilySearch remains essential for enterprise AI to prevent hallucination and access informationROI measurement focuses on cost reduction, analyst efficiency, and measurable business outcomesKey takeaway: invest in data structure early, context is king, AI is just softwareParticipants:Sebastien Paquet - Vice President of AI Strategy, CoveoRajan Koo - CTO, DTEX SystemsPatrick King - Head of Data, Honeycomb.ioKP Bhat - Sr Solutions Architecture Leader- Analytics & AI, Amazon Web ServicesFurther Links:Coveo: Website – LinkedIn – AWS MarketplaceDTEX Systems: Website – LinkedIn – AWS MarketplaceHoneycomb.io: Website – LinkedIn – AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep141: Securing Identities in the Age of AI Agents with Bhwana Singh, CTO of Okta

AWS for Software Companies Podcast

Play Episode Listen Later Sep 5, 2025 27:47


Okta's CTO Bhawna Singh discusses AI adoption, innovation and the four critical identity patterns needed to build the trust that accelerates AI implementation.Topics Include:AI innovation races ahead while adoption lags due to trust and security concernsResearch shows 82% plan AI deployment but 61% of customers demand trust firstAI coding tools dramatically reduce development time, accelerating software delivery cyclesAI interaction evolved from ChatGPT conversations to autonomous headless agents working independentlyFuture envisions millions of agents making decisions and communicating without human oversightComplex data relationships emerge as agents access multiple dynamic sources simultaneouslyTrust fundamentally starts with identity - the foundation for all AI securityFour critical identity patterns needed: authentication, API security, user confirmation, and authorizationAuthentication ensures legitimate agents while token vaults enable secure agent-to-agent communicationAsynchronous user approval prevents rogue decisions like the recent database deletion incidentIndustry standards like MCP protocol establish minimum security guardrails for interoperabilityTrust accelerates AI adoption through security, accountability, and collaborative standard-building effortsParticipants:Bhawna Singh – CTO, Customer Identity, OktaSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep140: Architecting Agentic AI Systems - Technical Insights for ISVs with Anyscale, CrewAI and Encrypt AI

AWS for Software Companies Podcast

Play Episode Listen Later Sep 3, 2025 44:28


A panel discussion with AI industry leaders revealing how enterprises are scaling AI today, with predictions on coming breakthroughs for AI and the impact on Fortune 500 companies and beyond.Topics Include:Three technical leaders discuss production challenges: security, interoperability, and scaling agentic systemsPanelists represent Enkrypt (security), Anyscale (infrastructure), and CrewAI (agent orchestration platforms)Industry moving from flashy demos to dependable agents with real business outcomesBreakthrough examples include 70-page IRS form processing and multimodal workflow automationMultimodal data integration becoming crucial - incorporating video, audio, screenshots into decisionsLess than 10% of future applications expected to be text-onlyCompanies shifting from experimenting with individual models to deploying agent networksNeed for governance frameworks as enterprises scale to hundreds of agentsGrowing software stack complexity requires specialized infrastructure between applications and GPUsSecurity teams need centralized visibility across fragmented agent deployments across enterprisesExisting industry regulations apply to AI services - no special AI laws neededInteroperability standards debate: MCP gaining adoption while A2A seems premature solutionMCP shows higher API reliability than OpenAI tool calling for implementationsMultimodal systems more vulnerable to attacks but value proposition too high ignoreFortune 500 company automated price operations approval process using 630 brands data87% of enterprise customers deploy agents in private VPCs or on-premises infrastructureSpecialized AI systems needed to oversee other agents at machine speed scalesCost optimization through model specialization rather than always using most powerful modelsFuture learning may happen through context/prompting rather than traditional weight fine-tuningPredictions include AI meeting moderators and agents working autonomously for hoursParticipants:Robert Nishihara - Co-founder, AnyscaleJoão Moura - CEO, CrewAISahil Agarwal - Co-Founder & CEO, Enkrypt AIJillian D'Arcy - Sr. ISV Sales Leader, Amazon Web ServicesFurther Links:Anyscale – Website | LinkedIn | AWS MarketplaceCrewAI - Website | LinkedIn | AWS MarketplaceEnkrypt AI - Website | LinkedIn | AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Pipeliners Podcast
Episode 404: Combining Gamification and Generative AI to Improve Training (with Survey) with Clint Bodungen

Pipeliners Podcast

Play Episode Listen Later Sep 2, 2025 42:07


In this episode of the Pipeliners Podcast, we revisit our conversation with Clint Bodungen of ThreatGEN. The discussion focuses on the application of gamification and generative AI in professional training, specifically for enhancing cybersecurity and incident response exercises. The episode also explores a PHMSA-sponsored R&D project that is adapting these advanced technologies for the unique operational needs of the pipeline industry, highlighting the development of AI-driven, multiplayer training environments.   Visit PipelinePodcastNetwork.com for a full episode transcript, as well as detailed show notes with relevant links and insider term definitions.

Oracle University Podcast
The AI Workflow

Oracle University Podcast

Play Episode Listen Later Sep 2, 2025 22:08


Join Lois Houston and Nikita Abraham as they chat with Yunus Mohammed, a Principal Instructor at Oracle University, about the key stages of AI model development. From gathering and preparing data to selecting, training, and deploying models, learn how each phase impacts AI's real-world effectiveness. The discussion also highlights why monitoring AI performance and addressing evolving challenges are critical for long-term success.   AI for You: https://mylearn.oracle.com/ou/course/ai-for-you/152601/252500   Oracle University Learning Community: https://education.oracle.com/ou-community   LinkedIn: https://www.linkedin.com/showcase/oracle-university/   X: https://x.com/Oracle_Edu   Special thanks to Arijit Ghosh, David Wright, Kris-Ann Nansen, Radhika Banka, and the OU Studio Team for helping us create this episode.   --------------------------------------------------------------   Episode Transcript: 00:00 Welcome to the Oracle University Podcast, the first stop on your cloud journey. During this series of informative podcasts, we'll bring you foundational training on the most popular Oracle technologies. Let's get started! 00:25 Lois: Welcome to the Oracle University Podcast! I'm Lois Houston, Director of Innovation Programs with Oracle University, and with me is Nikita Abraham, Team Lead: Editorial Services. Nikita: Hey everyone! In our last episode, we spoke about generative AI and gen AI agents. Today, we're going to look at the key stages in a typical AI workflow. We'll also discuss how data quality, feedback loops, and business goals influence AI success. With us today is Yunus Mohammed, a Principal Instructor at Oracle University.  01:00 Lois: Hi Yunus! We're excited to have you here! Can you walk us through the various steps in developing and deploying an AI model?  Yunus: The first point is the collect data. We gather relevant data, either historical or real time. Like customer transactions, support tickets, survey feedbacks, or sensor logs. A travel company, for example, can collect past booking data to predict future demand. So, data is the most crucial and the important component for building your AI models. But it's not just the data. You need to prepare the data. In the prepared data process, we clean, organize, and label the data. AI can't learn from messy spreadsheets. We try to make the data more understandable and organized, like removing duplicates, filling missing values in the data with some default values or formatting dates. All these comes under organization of the data and give a label to the data, so that the data becomes more supervised. After preparing the data, I go for selecting the model to train. So now, we pick what type of model fits your goals. It can be a traditional ML model or a deep learning network model, or it can be a generative model. The model is chosen based on the business problems and the data we have. So, we train the model using the prepared data, so it can learn the patterns of the data. Then after the model is trained, I need to evaluate the model. You check how well the model performs. Is it accurate? Is it fair? The metrics of the evaluation will vary based on the goal that you're trying to reach. If your model misclassifies emails as spam and it is doing it very much often, then it is not ready. So I need to train it further. So I need to train it to a level when it identifies the official mail as official mail and spam mail as spam mail accurately.  After evaluating and making sure your model is perfectly fitting, you go for the next step, which is called the deploy model. Once we are happy, we put it into the real world, like into a CRM, or a web application, or an API. So, I can configure that with an API, which is application programming interface, or I add it to a CRM, Customer Relationship Management, or I add it to a web application that I've got. Like for example, a chatbot becomes available on your company's website, and the chatbot might be using a generative AI model. Once I have deployed the model and it is working fine, I need to keep track of this model, how it is working, and need to monitor and improve whenever needed. So I go for a stage, which is called as monitor and improve. So AI isn't set in and forget it. So over time, there are lot of changes that is happening to the data. So we monitor performance and retrain when needed. An e-commerce recommendation model needs updates as there might be trends which are shifting.  So the end user finally sees the results after all the processes. A better product, or a smarter service, or a faster decision-making model, if we do this right. That is, if we process the flow perfectly, they may not even realize AI is behind it to give them the accurate results.  04:59 Nikita: Got it. So, everything in AI begins with data. But what are the different types of data used in AI development?  Yunus: We work with three main types of data: structured, unstructured, and semi-structured. Structured data is like a clean set of tables in Excel or databases, which consists of rows and columns with clear and consistent data information. Unstructured is messy data, like your email or customer calls that records videos or social media posts, so they all comes under unstructured data.  Semi-structured data is things like logs on XML files or JSON files. Not quite neat but not entirely messy either. So they are, they are termed semi-structured. So structured, unstructured, and then you've got the semi-structured. 05:58 Nikita: Ok… and how do the data needs vary for different AI approaches?  Yunus: Machine learning often needs labeled data. Like a bank might feed past transactions labeled as fraud or not fraud to train a fraud detection model. But machine learning also includes unsupervised learning, like clustering customer spending behavior. Here, no labels are needed. In deep learning, it needs a lot of data, usually unstructured, like thousands of loan documents, call recordings, or scan checks. These are fed into the models and the neural networks to detect and complex patterns. Data science focus on insights rather than the predictions. So a data scientist at the bank might use customer relationship management exports and customer demographies to analyze which age group prefers credit cards over the loans. Then we have got generative AI that thrives on diverse, unstructured internet scalable data. Like it is getting data from books, code, images, chat logs. So these models, like ChatGPT, are trained to generate responses or mimic the styles and synthesize content. So generative AI can power a banking virtual assistant trained on chat logs and frequently asked questions to answer customer queries 24/7. 07:35 Lois: What are the challenges when dealing with data?  Yunus: Data isn't just about having enough. We must also think about quality. Is it accurate and relevant? Volume. Do we have enough for the model to learn from? And is my data consisting of any kind of unfairly defined structures, like rejecting more loan applications from a certain zip code, which actually gives you a bias of data? And also the privacy. Are we handling personal data responsibly or not? Especially data which is critical or which is regulated, like the banking sector or health data of the patients. Before building anything smart, we must start smart.  08:23 Lois: So, we've established that collecting the right data is non-negotiable for success. Then comes preparing it, right?  Yunus: This is arguably the most important part of any AI or data science project. Clean data leads to reliable predictions. Imagine you have a column for age, and someone accidentally entered an age of like 999. That's likely a data entry error. Or maybe a few rows have missing ages. So we either fix, remove, or impute such issues. This step ensures our model isn't misled by incorrect values. Dates are often stored in different formats. For instance, a date, can be stored as the month and the day values, or it can be stored in some places as day first and month next. We want to bring everything into a consistent, usable format. This process is called as transformation. The machine learning models can get confused if one feature, like example the income ranges from 10,000 to 100,000, and another, like the number of kids, range from 0 to 5. So we normalize or scale values to bring them to a similar range, say 0 or 1. So we actually put it as yes or no options. So models don't understand words like small, medium, or large. We convert them into numbers using encoding. One simple way is assigning 1, 2, and 3 respectively. And then you have got removing stop words like the punctuations, et cetera, and break the sentence into smaller meaningful units called as tokens. This is actually used for generative AI tasks. In deep learning, especially for Gen AI, image or audio inputs must be of uniform size and format.  10:31 Lois: And does each AI system have a different way of preparing data?  Yunus: For machine learning ML, focus is on cleaning, encoding, and scaling. Deep learning needs resizing and normalization for text and images. Data science, about reshaping, aggregating, and getting it ready for insights. The generative AI needs special preparation like chunking, tokenizing large documents, or compressing images. 11:06 Oracle University's Race to Certification 2025 is your ticket to free training and certification in today's hottest tech. Whether you're starting with Artificial Intelligence, Oracle Cloud Infrastructure, Multicloud, or Oracle Data Platform, this challenge covers it all! Learn more about your chance to win prizes and see your name on the Leaderboard by visiting education.oracle.com/race-to-certification-2025. That's education.oracle.com/race-to-certification-2025. 11:50 Nikita: Welcome back! Yunus, how does a user choose the right model to solve their business problem?  Yunus: Just like a business uses different dashboards for marketing versus finance, in AI, we use different model types, depending on what we are trying to solve. Like classification is choosing a category. Real-world example can be whether the email is a spam or not. Use in fraud detection, medical diagnosis, et cetera. So what you do is you classify that particular data and then accurately access that classification of data. Regression, which is used for predicting a number, like, what will be the price of a house next month? Or it can be a useful in common forecasting sales demands or on the cost. Clustering, things without labels. So real-world examples can be segmenting customers based on behavior for targeted marketing. It helps discovering hidden patterns in large data sets.  Generation, that is creating new content. So AI writing product description or generating images can be a real-world example for this. And it can be used in a concept of generative AI models like ChatGPT or Dall-E, which operates on the generative AI principles. 13:16 Nikita: And how do you train a model? Yunus: We feed it with data in small chunks or batches and then compare its guesses to the correct values, adjusting its thinking like weights to improve next time, and the cycle repeats until the model gets good at making predictions. So if you're building a fraud detection system, ML may be enough. If you want to analyze medical images, you will need deep learning. If you're building a chatbot, go for a generative model like the LLM. And for all of these use cases, you need to select and train the applicable models as and when appropriate. 14:04 Lois: OK, now that the model's been trained, what else needs to happen before it can be deployed? Yunus: Evaluate the model, assess a model's accuracy, reliability, and real-world usefulness before it's put to work. That is, how often is the model right? Does it consistently perform well? Is it practical in the real world to use this model or not? Because if I have bad predictions, doesn't just look bad, it can lead to costly business mistakes. Think of recommending the wrong product to a customer or misidentifying a financial risk.  So what we do here is we start with splitting the data into two parts. So we train the data by training data. And this is like teaching the model. And then we have got the testing data. This is actually used for checking how well the model has learned. So once trained, the model makes predictions. We compare the predictions to the actual answers, just like checking your answer after a quiz. We try to go in for tailored evaluation based on AI types. Like machine learning, we care about accuracy in prediction. Deep learning is about fitting complex data like voice or images, where the model repeatedly sees examples and tunes itself to reduce errors. Data science, we look for patterns and insights, such as which features will matter. In generative AI, we judge by output quality. Is it coherent, useful, and is it natural?  The model improves with the accuracy and the number of epochs the training has been done on.  15:59 Nikita: So, after all that, we finally come to deploying the model… Yunus: Deploying a model means we are integrating it into our actual business system. So it can start making decisions, automating tasks, or supporting customer experiences in real time. Think of it like this. Training is teaching the model. Evaluating is testing it. And deployment is giving it a job.  The model needs a home either in the cloud or inside your company's own servers. Think of it like putting the AI in place where it can be reached by other tools. Exposed via API or embedded in an app, or you can say application, this is how the AI becomes usable.  Then, we have got the concept of receives live data and returns predictions. So receives live data and returns prediction is when the model listens to real-time inputs like a user typing, or user trying to search or click or making a transaction, and then instantly, your AI responds with a recommendation, decisions, or results. Deploying the model isn't the end of the story. It is just the beginning of the AI's real-world journey. Models may work well on day one, but things change. Customer behavior might shift. New products get introduced in the market. Economic conditions might evolve, like the era of COVID, where the demand shifted and the economical conditions actually changed. 17:48 Lois: Then it's about monitoring and improving the model to keep things reliable over time. Yunus: The monitor and improve loop is a continuous process that ensures an AI model remains accurate, fair, and effective after deployment. The live predictions, the model is running in real time, making decisions or recommendations. The monitor performance are those predictions still accurate and helpful. Is latency acceptable? This is where we track metrics, user feedbacks, and operational impact. Then, we go for detect issues, like accuracy is declining, are responses feeling biased, are customers dropping off due to long response times? And the next step will be to reframe or update the model. So we add fresh data, tweak the logic, or even use better architectures to deploy the uploaded model, and the new version replaces the old one and the cycle continues again. 18:58 Lois: And are there challenges during this step? Yunus: The common issues, which are related to monitor and improve consist of model drift, bias, and latency of failures. In model drift, the model becomes less accurate as the environment changes. Or bias, the model may favor or penalize certain groups unfairly. Latency or failures, if the model is too slow or fails unpredictably, it disrupts the user experience. Let's take the loan approvals. In loan approvals, if we notice an unusually high rejection rate due to model bias, we might retrain the model with more diverse or balanced data. For a chatbot, we watch for customer satisfaction, which might arise due to model failure and fine-tune the responses for the model. So in forecasting demand, if the predictions no longer match real trends, say post-pandemic, due to the model drift, we update the model with fresh data.  20:11 Nikita: Thanks for that, Yunus. Any final thoughts before we let you go? Yunus: No matter how advanced your model is, its effectiveness depends on the quality of the data you feed it. That means, the data needs to be clean, structured, and relevant. It should map itself to the problem you're solving. If the foundation is weak, the results will be also. So data preparation is not just a technical step, it is a business critical stage. Once deployed, AI systems must be monitored continuously, and you need to watch for drops in performance for any bias being generated or outdated logic, and improve the model with new data or refinements. That's what makes AI reliable, ethical, and sustainable in the long run. 21:09 Nikita: Yunus, thank you for this really insightful session. If you're interested in learning more about the topics we discussed today, go to mylearn.oracle.com and search for the AI for You course.  Lois: That's right. You'll find skill checks to help you assess your understanding of these concepts. In our next episode, we'll discuss the idea of buy versus build in the context of AI. Until then, this is Lois Houston… Nikita: And Nikita Abraham, signing off! 21:39 That's all for this episode of the Oracle University Podcast. If you enjoyed listening, please click Subscribe to get all the latest episodes. We'd also love it if you would take a moment to rate and review us on your podcast app. See you again on the next episode of the Oracle University Podcast.

AWS for Software Companies Podcast
Ep139: Human-in-the-Loop by Design: Building AI Systems Responsibly

AWS for Software Companies Podcast

Play Episode Listen Later Sep 1, 2025 39:40


AI executives from Archer, Demandbase and Highspot and AWS reveal how they're tackling AI's biggest challenges—from securing data, managing regulatory changes and keeping humans in the loop.Topics Include:Three AI leaders introduce their companies: Archer, Demandbase and Highspot's approaches to enterprise AIDemandbase's data strategy: Customer data stays isolated, shared data requires consent, public sources fuel trainingGeographic complexity: AI compliance varies dramatically between Germany, US, Canada, and California regulationsHighSpot tackles sales bias: Granular questions replace generic assessments for more accurate rep evaluationsSBI framework applied to AI: Specific behavioral observations create better, more actionable sales coachingAI transparency through citations: Timestamped evidence lets managers verify AI feedback and catch hallucinationsArcher handles 20-30K monthly regulations: AI helps enterprises manage overwhelming compliance requirements at scaleTwo compliance types explained: Operational (common across companies) versus business-specific regulatory requirementsEU AI Act adoption: US companies embracing European framework for responsible AI governanceHuman oversight becomes mandatory: Expert-in-the-loop reviews ensure AI decisions remain correctable and auditableThe bigger AI risk: Companies face greater danger from AI inaction than AI adoptionAgentic AI security challenges: Data layers must enforce permissions before AI access, not afterAI agents need identity management: Same access controls apply whether human clicks or AI actsHuman oversight in high stakes: Chief compliance officers demand transparency and correction capabilitiesFuture challenge identified: 80% of enterprise data behind firewalls remains invisible to AI modelsParticipants:Kayvan Alikhani - Global Head of Engineering- Emerging Solutions, Archer Integrated Risk ManagementUmberto Milletti - Chief R&D Officer, DemandbaseOliver Sharp - Co-Founder & Chief AI Officer, HighspotBrian Shadpour - General Manager, Security, Amazon Web ServicesFurther Links:Archer Integrated Risk Management: Website – LinkedIn – AWS MarketplaceDemandbase: Website – LinkedIn – AWS MarketplaceHighspot: Website – LinkedIn – AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep138: The Future of Agentic AI – Challenges and Opportunities with Rob McGrorty

AWS for Software Companies Podcast

Play Episode Listen Later Aug 29, 2025 31:22


In a fascinating discussion, Rob McGrorty, Product Leader of Agents at Amazon AGI Lab, reveals how rapidly AI agents are evolving with corporate adoption exploding as companies race to deploy production agents and the challenges and advantages they're experiencing.Topics Include:GenAI adoption outpaces all previous tech waves, growing faster than computers or internetEarly adopters tackle complex tasks while newcomers still use basic text manipulation featuresAI models double their single-call task capabilities every seven months, exponentially increasing powerAccelerating progress makes yesterday's magic mundane, unlocking mass creativity and customer demandAgents represent natural evolution: chatbots answered questions, now agents autonomously accomplish tasksAmazon's browser agent finds apartments, maps distances, ranks options using multiple transit modesCorporate adoption exploded: 33% piloting agents in 2024, 67% moving to production nowTwo main agent types today: API calling with tool use, browser automationCurrent applications mirror "RPA 2.0" - form filling, data extraction, website QA testingFuture brings multi-agent systems, self-directing loops, and agent-to-agent negotiation scenariosMajor challenges: data privacy, oversight protocols, error responsibility, and ecosystem sustainabilityTechnical hurdles include real-time accuracy measurement, latency issues, and quality assurance frameworksParticipants:Rob McGrorty – Product Leader, Agents at Amazon AGI LabSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep137: AI Without Borders - Extending analyst capabilities across the modern SOC

AWS for Software Companies Podcast

Play Episode Listen Later Aug 27, 2025 31:09


Gagan Singh of Elastic discuses how agentic AI systems reduce analyst burnout by automatically triaging security alerts, resulting in measurable ROI for organizationsTopics Include:AI breaks security silos between teams, data, and tools in SOCsAttackers gain system access; SOC teams have only 40 minutes to detect/containAlert overload causes analyst burnout; thousands of low-value alerts overwhelm teams dailyAI inevitable for SOCs to process data, separate false positives from real threatsAgentic systems understand environment, reason through problems, take action without hand-holdingAttack discovery capability reduces hundreds of alerts to 3-4 prioritized threat discoveriesAI provides ROI metrics: processed alerts, filtered noise, hours saved for organizationsRAG (Retrieval Augmented Generation) prevents hallucination by adding enterprise context to LLMsAWS integration uses SageMaker, Bedrock, Anthropic models with Elasticsearch vector database capabilitiesEnd-to-end LLM observability tracks costs, tokens, invocations, errors, and performance bottlenecksJunior analysts detect nation-state attacks; teams shift from reactive to proactive securityFuture requires balancing costs, data richness, sovereignty, model choice, human-machine collaborationParticipants:Gagan Singh – Vice President Product Marketing, ElasticAdditional Links:Elastic – LinkedIn - Website – AWS Marketplace See how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Oracle University Podcast
Core AI Concepts – Part 3

Oracle University Podcast

Play Episode Listen Later Aug 26, 2025 23:02


Join hosts Lois Houston and Nikita Abraham, along with Principal AI/ML Instructor Himanshu Raj, as they discuss the transformative world of Generative AI. Together, they uncover the ways in which generative AI agents are changing the way we interact with technology, automating tasks and delivering new possibilities.   AI for You: https://mylearn.oracle.com/ou/course/ai-for-you/152601/252500   Oracle University Learning Community: https://education.oracle.com/ou-community   LinkedIn: https://www.linkedin.com/showcase/oracle-university/   X: https://x.com/Oracle_Edu   Special thanks to Arijit Ghosh, David Wright, Kris-Ann Nansen, Radhika Banka, and the OU Studio Team for helping us create this episode. ------------------------------------------------------- Episode Transcript: 00:00 Welcome to the Oracle University Podcast, the first stop on your cloud journey. During this series of informative podcasts, we'll bring you foundational training on the most popular Oracle technologies. Let's get started! 00:25 Lois: Welcome to the Oracle University Podcast! I'm Lois Houston, Director of Innovation Programs with Oracle University, and with me is Nikita Abraham, Team Lead of Editorial Services.   Nikita: Hi everyone! Last week was Part 2 of our conversation on core AI concepts, where we went over the basics of data science. In Part 3 today, we'll look at generative AI and gen AI agents in detail. To help us with that, we have Himanshu Raj, Principal AI/ML Instructor. Hi Himanshu, what's the difference between traditional AI and generative AI?  01:01 Himanshu: So until now, when we talked about artificial intelligence, we usually meant models that could analyze information and make decisions based on it, like a judge who looks at evidence and gives a verdict. And that's what we call traditional AI that's focused on analysis, classification, and prediction.  But with generative AI, something remarkable happens. Generative AI does not just evaluate. It creates. It's more like a storyteller who uses knowledge from the past to imagine and build something brand new. For example, instead of just detecting if an email is spam, generative AI could write an entirely new email for you.  Another example, traditional AI might predict what a photo contains. Generative AI, on the other hand, creates a brand-new photo based on description. Generative AI refers to artificial intelligence models that can create entirely new content, such as text, images, music, code, or video that resembles human-made work.  Instead of simple analyzing or predicting, generative AI produces something original that resembles what a human might create.   02:16 Lois: How did traditional AI progress to the generative AI we know today?  Himanshu: First, we will look at small supervised learning. So in early days, AI models were trained on small labeled data sets. For example, we could train a model with a few thousand emails labeled spam or not spam. The model would learn simple decision boundaries. If email contains, "congratulations," it might be spam. This was efficient for a straightforward task, but it struggled with anything more complex.  Then, comes the large supervised learning. As the internet exploded, massive data sets became available, so millions of images, billions of text snippets, and models got better because they had much more data and stronger compute power and thanks to advances, like GPUs, and cloud computing, for example, training a model on millions of product reviews to predict customer sentiment, positive or negative, or to classify thousands of images in cars, dogs, planes, etc.  Models became more sophisticated, capturing deeper patterns rather than simple rules. And then, generative AI came into the picture, and we eventually reached a point where instead of just classifying or predicting, models could generate entirely new content.  Generative AI models like ChatGPT or GitHub Copilot are trained on enormous data sets, not to simply answer a yes or no, but to create outputs that look and feel like human made. Instead of judging the spam or sentiment, now the model can write an article, compose a song, or paint a picture, or generate new software code.  03:55 Nikita: Himanshu, what motivated this sort of progression?   Himanshu: Because of the three reasons. First one, data, we had way more of it thanks to the internet, smartphones, and social media. Second is compute. Graphics cards, GPUs, parallel computing, and cloud systems made it cheap and fast to train giant models.  And third, and most important is ambition. Humans always wanted machines not just to judge existing data, but to create new knowledge, art, and ideas.   04:25 Lois: So, what's happening behind the scenes? How is gen AI making these things happen?  Himanshu: Generative AI is about creating entirely new things across different domains. On one side, we have large language models or LLMs.  They are masters of generating text conversations, stories, emails, and even code. And on the other side, we have diffusion models. They are the creative artists of AI, turning text prompts into detailed images, paintings, or even videos.  And these two together are like two different specialists. The LLM acts like a brain that understands and talks, and the diffusion model acts like an artist that paints based on the instructions. And when we connect these spaces together, we create something called multimodal AI, systems that can take in text and produce images, audio, or other media, opening a whole new range of possibilities.  It can not only take the text, but also deal in different media options. So today when we say ChatGPT or Gemini, they can generate images, and it's not just one model doing everything. These are specialized systems working together behind the scenes.  05:38 Lois: You mentioned large language models and how they power text-based gen AI, so let's talk more about them. Himanshu, what is an LLM and how does it work?  Himanshu: So it's a probabilistic model of text, which means, it tries to predict what word is most likely to come next based on what came before.  This ability to predict one word at a time intelligently is what builds full sentences, paragraphs, and even stories.  06:06 Nikita: But what's large about this? Why's it called a large language model?   Himanshu: It simply means the model has lots and lots of parameters. And think of parameters as adjustable dials the model fine tuned during learning.  There is no strict rule, but today, large models can have billions or even trillions of these parameters. And the more the parameters, more complex patterns, the model can understand and can generate a language better, more like human.  06:37 Nikita: Ok… and image-based generative AI is powered by diffusion models, right? How do they work?  Himanshu: Diffusion models start with something that looks like pure random noise.  Imagine static on an old TV screen. No meaningful image at all. From there, the model carefully removes noise step by step to create something more meaningful and think of it like sculpting a statue. You start with a rough block of stone and slowly, carefully you chisel away to reveal a beautiful sculpture hidden inside.  And in each step of this process, the AI is making an educated guess based on everything it has learned from millions of real images. It's trying to predict.   07:24 Stay current by taking the 2025 Oracle Fusion Cloud Applications Delta Certifications. This is your chance to demonstrate your understanding of the latest features and prove your expertise by obtaining a globally recognized certification, all for free! Discover the certification paths, use the resources on MyLearn to prepare, and future-proof your skills. Get started now at mylearn.oracle.com.  07:53 Nikita: Welcome back! Himanshu, for most of us, our experience with generative AI is with text-based tools like ChatGPT. But I'm sure the uses go far beyond that, right? Can you walk us through some of them?  Himanshu: First one is text generation. So we can talk about chatbots, which are now capable of handling nuanced customer queries in banking travel and retail, saving companies hours of support time. Think of a bank chatbot helping a customer understand mortgage options or virtual HR Assistant in a large company, handling leave request. You can have embedding models which powers smart search systems.  Instead of searching by keywords, businesses can now search by meaning. For instance, a legal firm can search cases about contract violations in tech and get semantically relevant results, even if those exact words are not used in the documents.  The third one, for example, code generation, tools like GitHub Copilot help developers write boilerplate or even functional code, accelerating software development, especially in routine or repetitive tasks. Imagine writing a waveform with just a few prompts.  The second application, is image generation. So first obvious use is art. So designers and marketers can generate creative concepts instantly. Say, you need illustrations for a campaign on future cities. Generative AI can produce dozens of stylized visuals in minutes.  For design, interior designers or architects use it to visualize room layouts or design ideas even before a blueprint is finalized. And realistic images, retail companies generate images of people wearing their clothing items without needing real models or photoshoots, and this reduces the cost and increase the personalization.  Third application is multimodal systems, and these are combined systems that take one kind of input or a combination of different inputs and produce different kind of outputs, or can even combine various kinds, be it text image in both input and output.  Text to image It's being used in e-commerce, movie concept art, and educational content creation. For text to video, this is still in early days, but imagine creating a product explainer video just by typing out the script. Marketing teams love this for quick turnarounds. And the last one is text to audio.  Tools like ElevenLabs can convert text into realistic, human like voiceovers useful in training modules, audiobooks, and accessibility apps. So generative AI is no longer just a technical tool. It's becoming a creative copilot across departments, whether it's marketing, design, product support, and even operations.  10:42 Lois: That's great! So, we've established that generative AI is pretty powerful. But what kind of risks does it pose for businesses and society in general?  Himanshu: The first one is deepfakes. Generative AI can create fake but highly realistic media, video, audios or even faces that look and sound authentic.  Imagine a fake video of a political leader announcing a policy, they never approved. This could cause mass confusion or even impact elections. In case of business, deepfakes can be also used in scams where a CEO's voice is faked to approve fraudulent transactions.  Number two, bias, if AI is trained on biased historical data, it can reinforce stereotypes even when unintended. For example, a hiring AI system that favors male candidates over equally qualified women because of historical data was biased.  And this bias can expose companies to discrimination, lawsuits, brand damage and ethical concerns. Number three is hallucinations. So sometimes AI system confidently generate information that is completely wrong without realizing it.   Sometimes you ask a chatbot for a legal case summary, and it gives you a very convincing but entirely made up court ruling. In case of business impact, sectors like health care, finance, or law hallucinations can or could have serious or even dangerous consequences if not caught.  The fourth one is copyright and IP issues, generative AI creates new content, but often, based on material it was trained on. Who owns a new work? A real life example could be where an artist finds their unique style was copied by an AI that was trained on their paintings without permission.  In case of a business impact, companies using AI-generated content for marketing, branding or product designs must watch for legal gray areas around copyright and intellectual properties. So generative AI is not just a technology conversation, it's a responsibility conversation. Businesses must innovate and protect.  Creativity and caution must go together.   12:50 Nikita: Let's move on to generative AI agents. How is a generative AI agent different from just a chatbot or a basic AI tool?  Himanshu: So think of it like a smart assistant, not just answering your questions, but also taking actions on your behalf. So you don't just ask, what's the best flight to Vegas? Instead, you tell the agent, book me a flight to Vegas and a room at the Hilton. And it goes ahead, understands that, finds the options, connects to the booking tools, and gets it done.   So act on your behalf using goals, context, and tools, often with a degree of autonomy. Goals, are user defined outcomes. Example, I want to fly to Vegas and stay at Hilton. Context, this includes preferences history, constraints like economy class only or don't book for Mondays.  Tools could be APIs, databases, or services it can call, such as a travel API or a company calendar. And together, they let the agent reason, plan, and act.   14:02 Nikita: How does a gen AI agent work under the hood?  Himanshu: So usually, they go through four stages. First, one is understands and interprets your request like natural language understanding. Second, figure out what needs to be done, in this case flight booking plus hotel search.  Third, retrieves data or connects to tools APIs if needed, such as Skyscanner, Expedia, or a Calendar. And fourth is takes action. That means confirming the booking and giving you a response like your travel is booked. Keep in mind not all gen AI agents are fully independent.  14:38 Lois: Himanshu, we've seen people use the terms generative AI agents and agentic AI interchangeably. What's the difference between the two?  Himanshu: Agentic AI is a broad umbrella. It refers to any AI system that can perceive, reason, plan, and act toward a goal and may improve and adapt over time.   Most gen AI agents are reactive, not proactive. On the other hand, agentic AI can plan ahead, anticipate problems, and can even adjust strategies.  So gen AI agents are often semi-autonomous. They act in predefined ways or with human approval. Agentic systems can range from low to full autonomy. For example, auto-GPT runs loops without user prompts and autonomous car decides routes and reactions.  Most gen AI agents can only make multiple steps if explicitly designed that way, like a step-by-step logic flows in LangChain. And in case of agentic AI, it can plan across multiple steps with evolving decisions.  On the memory and goal persistence, gen AI agents are typically stateless. That means they forget their goal unless you remind them. In case of agentic AI, these systems remember, adapt, and refine based on goal progression. For example, a warehouse robot optimizing delivery based on changing layouts.  Some generative AI agents are agentic, like auto GPT. They use LLMs to reason, plan, and act, but not all. And likewise not all agentic AIs are generative. For example, an autonomous car, which may use computer vision control systems and planning, but no generative models.  So agentic AI is a design philosophy or system behavior, which could be goal-driven, autonomous, and decision making. They can overlap, but as I said, not all generative AI agents are agentic, and not all agentic AI systems are generative.  16:39 Lois: What makes a generative AI agent actually work?  Himanshu: A gen AI agent isn't just about answering the question. It's about breaking down a user's goal, figuring out how to achieve it, and then executing that plan intelligently. These agents are built from five core components and each playing a critical role.  The first one is goal. So what is this agent trying to achieve? Think of this as the mission or intent. For example, if I tell the agent, help me organized a team meeting for Friday. So the goal in that case would be schedule a meeting.  Number 2, memory. What does it remember? So this is the agent's context awareness. Storing previous chats, preferences, or ongoing tasks. For example, if last week I said I prefer meetings in the afternoon or I have already shared my team's availability, the agent can reuse that. And without the memory, the agent behaves stateless like a typical chatbot that forgets context after every prompt.  Third is tools. What can it access? Agents aren't just smart, they are also connected. They can be given access to tools like calendars, CRMs, web APIs, spreadsheets, and so on.  The fourth one is planner. So how does it break down the goal? And this is where the reasoning happens. The planner breaks big goals into a step-by-step plans, for example checking team availability, drafting meeting invite, and then sending the invite. And then probably, will confirm the booking. Agents don't just guess. They reason and organize actions into a logical path.  And the fifth and final one is executor, who gets it done. And this is where the action takes place. The executor performs what the planner lays out. For example, calling APIs, sending message, booking reservations, and if planner is the architect, executor is the builder.   18:36 Nikita: And where are generative AI agents being used?  Himanshu: Generative AI agents aren't just abstract ideas, they are being used across business functions to eliminate repetitive work, improve consistency, and enable faster decision making. For marketing, a generative AI agent can search websites and social platforms to summarize competitor activity. They can draft content for newsletters or campaign briefs in your brand tone, and they can auto-generate email variations based on audience segment or engagement history.  For finance, a generative AI agent can auto-generate financial summaries and dashboards by pulling from ERP spreadsheets and BI tools. They can also draft variance analysis and budget reports tailored for different departments. They can scan regulations or policy documents to flag potential compliance risks or changes.  For sales, a generative AI agent can auto-draft personalized sales pitches based on customer behavior or past conversations. They can also log CRM entries automatically once submitting summary is generated. They can also generate battlecards or next-step recommendations based on the deal stage.  For human resource, a generative AI agent can pre-screen resumes based on job requirements. They can send interview invites and coordinate calendars. A common theme here is that generative AI agents help you scale your teams without scaling the headcount.   20:02 Nikita: Himanshu, let's talk about the capabilities and benefits of generative AI agents.  Himanshu: So generative AI agents are transforming how entire departments function. For example, in customer service, 24/7 AI agents handle first level queries, freeing humans for complex cases.  They also enhance the decision making. Agents can quickly analyze reports, summarize lengthy documents, or spot trends across data sets. For example, a finance agent reviewing Excel data can highlight cash flow anomalies or forecast trends faster than a team of analysts.  In case of personalization, the agents can deliver unique, tailored experiences without manual effort. For example, in marketing, agents generate personalized product emails based on each user's past behavior. For operational efficiency, they can reduce repetitive, low-value tasks. For example, an HR agent can screen hundreds of resumes, shortlist candidates, and auto-schedule interviews, saving HR team hours each week.  21:06 Lois: Ok. And what are the risks of using generative AI agents?  Himanshu: The first one is job displacement. Let's be honest, automation raises concerns. Roles involving repetitive tasks such as data entry, content sorting are at risk. In case of ethics and accountability, when an AI agent makes a mistake, who is responsible? For example, if an AI makes a biased hiring decision or gives incorrect medical guidance, businesses must ensure accountability and fairness.  For data privacy, agents often access sensitive data, for example employee records or customer history. If mishandled, it could lead to compliance violations. In case of hallucinations, agents may generate confident but incorrect outputs called hallucinations. This can often mislead users, especially in critical domains like health care, finance, or legal.  So generative AI agents aren't just tools, they are a force multiplier. But they need to be deployed thoughtfully with a human lens and strong guardrails. And that's how we ensure the benefits outweigh the risks.  22:10 Lois: Thank you so much, Himanshu, for educating us. We've had such a great time with you! If you want to learn more about the topics discussed today, head over to mylearn.oracle.com and get started on the AI for You course.  Nikita: Join us next week as we chat about AI workflows and tools. Until then, this is Nikita Abraham…  Lois: And Lois Houston signing off!  22:32 That's all for this episode of the Oracle University Podcast. If you enjoyed listening, please click Subscribe to get all the latest episodes. We'd also love it if you would take a moment to rate and review us on your podcast app. See you again on the next episode of the Oracle University Podcast.  

AWS for Software Companies Podcast
Ep136: Rapid7's Journey to an AI First Platform with AWS

AWS for Software Companies Podcast

Play Episode Listen Later Aug 25, 2025 25:26


Pete Rubio reveals how Rapid7 transformed to an AI-first platform that automates security investigations and accelerates results from hours to seconds.Topics Include:Pete Rubio introduces Rapid7's journey to becoming an AI-first cybersecurity platformCybersecurity teams overwhelmed by growing attack surfaces and constant alert fatigueCustomers needed faster response times, not just more alerts coming fasterLegacy tools created silos requiring manual triage that doesn't scale effectivelyAI must turn raw security data into real-time decisions humans can trustUnified data platform correlates infrastructure, applications, identity, and business context togetherAgentic AI automates investigative work, reducing analyst tasks from hours to secondsRapid7 evaluated multiple vendors, choosing AWS for performance, cost, and flexibilityNova models delivered unmatched performance for global scaling at controlled costsBedrock provided secure model deployment with governance and data privacy boundariesAWS partnership enabled co-development and rapid iteration beyond typical vendor relationshipsTransparent AI shows customers how models reach conclusions before automated actionsSOC analyst expertise continuously trains models with real-time security intelligenceGovernance frameworks and guardrails implemented from day one, not retrofitted laterFuture plans include customer AI integration and bring-your-own-model capabilitiesParticipants:Pete Rubio – Senior Vice President, Platform & Engineering, Rapid7Additional Links:Rapid 7 – LinkedIn - Website – AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep135: Petabytes and Milliseconds: How Panther scales Security Monitoring with Cloud-Native AI

AWS for Software Companies Podcast

Play Episode Listen Later Aug 22, 2025 10:49


Panther CEO William Lowe explains how integrating Amazon Bedrock AI into their security platform delivered 50% faster alert resolution for enterprise customers while maintaining the trust and control that security practitioners demand.Topics Include:Panther CEO explains how Amazon partnership accelerates security outcomes for customersCloud-native security platform delivers 100% visibility across enterprise environments at scaleCustomers like Dropbox and Coinbase successfully replaced Splunk with Panther's solutionPlatform processes petabytes monthly with impressive 2.3-minute average threat detection timeCritical gap identified: alert resolution still takes 8 hours despite fast detectionSecurity teams overwhelmed by growing attack surfaces and severe talent burnoutConstant context switching across tools creates inefficiency and organizational collaboration problemsAI integration with Amazon Bedrock designed to accelerate security team decision-makingFour trust principles: verifiable actions, secure design, human control, customer data ownershipResults show 50% faster alert triage; future includes Slack integration and automationParticipants:· William H Lowe – CEO, PantherSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep134: Prime Opportunities for ISVs by Leveraging Generative AI

AWS for Software Companies Podcast

Play Episode Listen Later Aug 20, 2025 30:43


AWS executives reveal how generative AI is fundamentally reshaping ISV business models, from pricing strategies to go-to-market approaches, and provide actionable insights for software companies navigating this transformation.Topics Include:Alayna Broaderson and Andy Perkins introduce AWS Infrastructure Partnerships and ISV SalesGenerative AI profoundly changing how ISVs build, deliver and market software productsTwo ISV categories emerging: established SaaS companies versus pure gen AI startupsLegacy SaaS firms struggle with infrastructure modernization and potential revenue cannibalizationPure gen AI companies face scaling challenges, reliability issues and cost optimizationRevenue models shifting from subscription-based to consumption-based pricing per token/prompt/taskFuture-proofing architecture critical as technology evolves rapidly like F-35 fighter jetsData becoming key differentiator, especially domain-specific datasets in healthcare and legalBalancing cost, accuracy, latency and customer experience creates complex optimization challengesMultiple specialized models replacing single solutions, with agentic AI accelerating this trendHuman capital challenges include retraining engineering teams and finding expensive AI talentSecurity, compliance and explainability now mandatory - no more black box solutionsEnterprise customers struggle with data organization and quantifying clear gen AI ROIISV pricing models evolving with tiered structures and targeted vertical use casesTraditional SaaS playbooks failing in generative AI landscape due to ROI uncertaintyPOC-based go-to-market with free trials and case study selling proving most effectivePricing strategies include AI gates, credit systems and separate SKUs for servicesCustomer trust requires proactive security messaging and auditable, transparent AI solutionsModular architecture enables evolution as new technologies emerge in fast-changing marketAWS positioning as ultimate gen AI toolkit partner with ISV collaboration opportunitiesParticipants:Alayna Broaderson - Sr Manager, Infrastructure Technology Partnership, Amazon Web ServicesAndy Perkins - General Manager, US ISV, Amazon Web ServicesSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep133: Enabling Better Customer Experiences with Amazon Q Index w/ PagerDuty and Zoom

AWS for Software Companies Podcast

Play Episode Listen Later Aug 18, 2025 23:10


Hear how PagerDuty and Zoom built successful AI products using Amazon Q-Index to solve real customer problems like incident response and meeting intelligence, while sharing practical lessons from their early adoption journey.Topics Include:David Gordon introduces AWS Q-Business partnerships with PagerDuty and ZoomMeet Everaldo Aguiar: PagerDuty's Applied AI leader with academia and enterprise backgroundPaul Magnaghi from Zoom brings AI platform scaling experience from SeattleQ-Business launched over a year ago as managed generative AI servicePlatform enables agentic experiences: content discovery, analysis, and process automationBuilt on AWS Bedrock with enterprise guardrails and data source integrationPartners wanted backend capabilities but preferred their own UI and modelsQ-Index provides vector database functionality for ISV partner integrationsEveraldo explains PagerDuty's evolution from traditional ML to generative AI solutionsHistorical challenges: alert fatigue, noise reduction using machine learning approachesNew gen AI opportunities: incident context, relevant data surfacing, automated postmortemsEngineering teams faced learning curve with agents and high-latency user experiencesPaul discusses Zoom's existing AI: virtual backgrounds and voice isolation technologyAI Companion strategy focused on simplicity during complex generative AI adoptionProblem identified: valuable meeting conversations disappear after Zoom calls endCustomer feedback revealed need for enterprise data integration beyond basic summariesGoal: combine unstructured conversations with structured enterprise data seamlesslyPagerDuty Advanced provides agentic AI for on-call engineers during incidentsQ-Index integration accesses internal documentation: Confluence pages, runbooks, proceduresDemo shows Slack integration pulling relevant incident response documentation automaticallyAccess control lists ensure users see only data they're authorized to accessZoom's AI companion panel enables real-time meeting questions and summariesExample use cases: decision tracking, incident analysis, action item identificationAdvice for starting: standardize practices and create internal development templatesSingle data access point reduces legal and security evaluation overheadCenter of excellence approach helps teams move quickly across product divisionsCut through generative AI buzzwords to focus on real user valueFederated AWS Bedrock architecture provides model choice and flexibility meeting customersCustomer trust alignment between Zoom conversations and AWS data handlingGetting started: PagerDuty Advance available now, Zoom AI free with paid add-onsParticipants:Everaldo Aguiar – Senior Engineering Manager, Applied AI, PagerDutyPaul Magnaghi – Head of AI & ISV Go To Market, ZoomDavid Gordon - Global Business Development, Amazon Q for Business. Amazon Web ServicesFurther Links:PagerDuty Website, LinkedIn & AWS MarketplaceZoom Website, LinkedIn & AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep132: Security vs Productivity – Winning the AI Arms-Race with Teleport and AWS

AWS for Software Companies Podcast

Play Episode Listen Later Aug 15, 2025 31:41


Teleport Co-Founder and CEO Ev Kontsevoy discusses the security vs productivity trade-off that plagues growing companies and how Teleport's trusted computing model protects against the exponential growth of cybersecurity threats.Topics Include:Teleport CEO explains how to make infrastructure "nearly unhackable" through trusted computingTraditional security vs productivity trade-off: high security kills team efficiencyCompanies buy every security solution but still get told they're at riskWhy "crown jewels" thinking fails: computers should protect everything at scaleModern infrastructure has too many access paths to enumerate and secureApple's PCC specification shows trusted computing working in real production environmentsAI revolutionizes both offensive and defensive cybersecurity capabilities for everyone80% of companies can't guarantee they've removed all ex-employee accessIdentity fragmentation across systems creates anonymous relationships and security gapsHuman error probability grows exponentially as companies scale in three dimensionsYour laptop already demonstrates trusted computing: seamless access without constant loginsApple ecosystem shows device trust at scale through secure enclavesAI agents need trusted identities just like humans and machinesAWS marketplace partnership accelerates deals and provides strategic account insightsHire someone who understands partnership dynamics before starting with AWSGenerative AI will make identity attacks cheaper and faster than everSecurity responsibility shifting from IT teams to platform engineering teamsTeleport's "steady state invariant": infrastructure locked down except during authorized workTemporary access granted through tickets, then automatically revoked after completionLegacy systems and IoT devices require extending trust models beyond cloud-nativeParticipants:Ev Kontsevoy – Co-Founder and CEO, TeleportFurther Links:Teleport WebsiteTeleport AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep131: Preventing Identity Theft at Scale: How DTEX Systems Detects and Disarms Insider Threats with Amazon Bedrock

AWS for Software Companies Podcast

Play Episode Listen Later Aug 13, 2025 15:08


Raj Koo, CTO of DTEX Systems, discusses how their enterprise-grade generative AI platform detects and disarms insider threats and enables them to stay ahead of evolving risks.Topics Include:Raj Koo, CTO of DTEX Systems, joins from Adelaide to discuss insider threat detectionDTEX evolved from Adelaide startup to Bay Area headquarters, serving Fortune 500 companiesCompany specializes in understanding human behavior and intention behind insider threatsMarket shifting beyond cyber indicators to focus on behavioral analysis and detectionRecent case: US citizen sold identity to North Korean DPRK IT workersForeign entities used stolen credentials to infiltrate American companies undetectedDTEX's behavioral detection systems helped identify this sophisticated identity theft operationGenerative AI becomes double-edged sword - used by both threat actors and defendersBad actors use AI for fake resumes and deepfake interviewsDTEX uses traditional machine learning for risk modeling, GenAI for analyst interpretationGoal is empowering security analysts to work faster, not replacing human expertiseAWS GenAI Innovation Center helped develop guardrails and usage boundaries for enterpriseChallenge: enterprises must follow rules while hackers operate without ethical constraintsDTEX gains advantage through proprietary datasets unavailable to public AI modelsAWS Bedrock partnership enables private, co-located language models for data securityPrivate preview launched February 2024 with AWS Innovation Center acceleration supportSoftware leaders should prioritize privacy-by-design from day one of GenAI adoptionFuture threat: information sharing shifts from files to AI-powered data queriesMonitoring who asks what questions of AI systems becomes critical security concernDTEX contributes to OpenSearch development while building vector databases for analysisParticipants:Rajan Koo – Chief Technology Officer, DTEX SystemsFurther Links:DTEX Systems WebsiteDTEX Systems AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Clownfish TV: Audio Edition
The Internet is DEAD and So is Twitch.

Clownfish TV: Audio Edition

Play Episode Listen Later Aug 12, 2025 15:58


Half of the internet is fake, bot traffic according to reports. Then we talk about how most of the top Twitch streamers were BUYING viewers to bilk advertisers. The dead internet is here. Watch this podcast episode on YouTube and all major podcast hosts including Spotify. CLOWNFISH TV is an independent, opinionated news and commentary podcast that covers Entertainment and Tech from a consumer's point of view. We talk about Gaming, Comics, Anime, TV, Movies, Animation and more. Hosted by Kneon and Geeky Sparkles. D/REZZED News covers Pixels, Pop Culture, and the Paranormal! We're an independent, opinionated entertainment news blog covering Video Games, Tech, Comics, Movies, Anime, High Strangeness, and more. As part of Clownfish TV, we strive to be balanced, based, and apolitical. Get more news, views and reviews on Clownfish TV News - https://news.clownfishtv.com/ On YouTube - https://www.youtube.com/c/ClownfishTV On Spotify - https://open.spotify.com/show/4Tu83D1NcCmh7K1zHIedvg On Apple Podcasts - https://podcasts.apple.com/us/podcast/clownfish-tv-audio-edition/id1726838629

AWS for Software Companies Podcast
Ep130: Agentic AI - Transforming Enterprise Technology with leaders from C3 AI, Resolve AI and Scale AI

AWS for Software Companies Podcast

Play Episode Listen Later Aug 11, 2025 30:39


Enterprise AI leaders from C3 AI, Resolve AI, and Scale AI reveal how Fortune 100 companies are successfully scaling agentic AI from pilots to production and share secrets for successful AI transformation.Topics Include:Panel introduces three AI leaders from Resolve AI, C3 AI, and Scale AIResolve AI builds autonomous site reliability engineers for production incident responseC3 AI provides full-stack platform for developing enterprise agentic AI workflowsScale AI helps Fortune 100 companies adopt agents with private data integrationMoving from AI pilots to production requires custom solutions, not shrink-wrap softwareSuccess demands working directly with customers to understand their specific workflowsAll enterprise AI solutions need well-curated access to internal data and resourcesSoftware engineering has permanently shifted to agentic coding with no going backAI agents rapidly improving in reasoning, tool use, and contextual understandingIndustry moving from simple co-pilots to agents solving complex multi-step problemsSpiros coins new concept: evolving from "systems of record" to "systems of knowledge"Democratized development platforms let enterprises declare their own agent workflowsSemantic business layers enable agents to understand domain-specific enterprise operationsTrust and observability remain major barriers to enterprise agent adoptionOversight layers essential for agents making longer-horizon autonomous business decisionsPerformance tracking and calibration systems needed like MLOps for reasoning chainsCEO-level top-down support required for successful AI transformation initiativesTraditional per-seat SaaS pricing models completely broken for agentic AI solutionsIndustry shifting toward outcome-based and work-completion pricing models insteadReal examples shared: agent collaboration in production engineering and sales automationParticipants:Nikhil Krishnan – SVP & Chief Technology Officer, Data Science, C3 AISpiros Xanthos – Founder and CEO, Resolve AIVijay Karunamurthy – Head of Engineering, Product and Design / Field Chief Technology Officer, Scale AIAndy Perkins – GM, US ISV Sales – Data, Analytics, GenAI, Amazon Web ServicesFurther Links:C3 – Website – AWS MarketplaceResolve AI – Website – AWS MarketplaceScale AI – Website – AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Shadow Warrior by Rajeev Srinivasan
Ep 173: Trump tariff wars: Seeing them in context for India

Shadow Warrior by Rajeev Srinivasan

Play Episode Listen Later Aug 10, 2025 27:23


A version of this essay has been published by firstpost.com at https://www.firstpost.com/opinion/shadow-warrior-from-crisis-to-advantage-how-india-can-outplay-the-trump-tariff-gambit-13923031.htmlA simple summary of the recent brouhaha about President Trump's imposition of 25% tariffs on India as well as his comment on India's ‘dead economy' is the following from Shakespeare's Macbeth: “full of sound and fury, signifying nothing”. Trump further imposed punitive tariffs totalling 50% on August 6th allegedly for India funding Russia's war machine via buying oil.As any negotiator knows, a good opening gambit is intended to set the stage for further parleys, so that you could arrive at a negotiated settlement that is acceptable to both parties. The opening gambit could well be a maximalist statement, or one's ‘dream outcome', the opposite of which is ‘the walkway point' beyond which you are simply not willing to make concessions. The usual outcome is somewhere in between these two positions or postures.Trump is both a tough negotiator, and prone to making broad statements from which he has no problem retreating later. It's down-and-dirty boardroom tactics that he's bringing to international trade. Therefore I think Indians don't need to get rattled. It's not the end of the world, and there will be climbdowns and adjustments. Think hard about the long term.I was on a panel discussion on this topic on TV just hours after Trump made his initial 25% announcement, and I mentioned an interplay between geo-politics and geo-economics. Trump is annoyed that his Ukraine-Russia play is not making much headway, and also that BRICS is making progress towards de-dollarization. India is caught in this crossfire (‘collateral damage') but the geo-economic facts on the ground are not favorable to Trump.I am in general agreement with Trump on his objectives of bringing manufacturing and investment back to the US, but I am not sure that he will succeed, and anyway his strong-arm tactics may backfire. I consider below what India should be prepared to do to turn adversity into opportunity.The anti-Thucydides Trap and the baleful influence of Whitehall on Deep StateWhat is remarkable, though, is that Trump 2.0 seems to be indistinguishable from the Deep State: I wondered last month if the Deep State had ‘turned' Trump. The main reason many people supported Trump in the first place was the damage the Deep State was wreaking on the US under the Obama-Biden regime. But it appears that the resourceful Deep State has now co-opted Trump for its agenda, and I can only speculate how.The net result is that there is the anti-Thucydides Trap: here is the incumbent power, the US, actively supporting the insurgent power, China, instead of suppressing it, as Graham Allison suggested as the historical pattern. It, in all fairness, did not start with Trump, but with Nixon in China in 1971. In 1985, the US trade deficit with China was $6 million. In 1986, $1.78 billion. In 1995, $35 billion.But it ballooned after China entered the WTO in 2001. $202 billion in 2005; $386 billion in 2022.In 2025, after threatening China with 150% tariffs, Trump retreated by postponing them; besides he has caved in to Chinese demands for Nvidia chips and for exemptions from Iran oil sanctions if I am not mistaken.All this can be explained by one word: leverage. China lured the US with the siren-song of the cost-leader ‘China price', tempting CEOs and Wall Street, who sleepwalked into surrender to the heft of the Chinese supply chain.Now China has cornered Trump via its monopoly over various things, the most obvious of which is rare earths. Trump really has no option but to give in to Chinese blackmail. That must make him furious: in addition to his inability to get Putin to listen to him, Xi is also ignoring him. Therefore, he will take out his frustrations on others, such as India, the EU, Japan, etc. Never mind that he's burning bridges with them.There's a Malayalam proverb that's relevant here: “angadiyil thottathinu ammayodu”. Meaning, you were humiliated in the marketplace, so you come home and take it out on your mother. This is quite likely what Trump is doing, because he believes India et al will not retaliate. In fact Japan and the EU did not retaliate, but gave in, also promising to invest large sums in the US. India could consider a different path: not active conflict, but not giving in either, because its equations with the US are different from those of the EU or Japan.Even the normally docile Japanese are beginning to notice.Beyond that, I suggested a couple of years ago that Deep State has a plan to enter into a condominium agreement with China, so that China gets Asia, and the US gets the Americas and the Pacific/Atlantic. This is exactly like the Vatican-brokered medieval division of the world between Spain and Portugal, and it probably will be equally bad for everyone else. And incidentally it makes the Quad infructuous, and deepens distrust of American motives.The Chinese are sure that they have achieved the condominium, or rather forced the Americans into it. Here is a headline from the Financial Express about their reaction to the tariffs: they are delighted that the principal obstacle in their quest for hegemony, a US-India military and economic alliance, is being blown up by Trump, and they lose no opportunity to deride India as not quite up to the mark, whereas they and the US have achieved a G2 detente.Two birds with one stone: gloat about the breakdown in the US-India relationship, and exhibit their racist disdain for India yet again.They laugh, but I bet India can do an end-run around them. As noted above, the G2 is a lot like the division of the world into Spanish and Portuguese spheres of influence in 1494. Well, that didn't end too well for either of them. They had their empires, which they looted for gold and slaves, but it made them fat, dumb and happy. The Dutch, English, and French capitalized on more dynamic economies, flexible colonial systems, and aggressive competition, overtaking the Iberian powers in global influence by the 17th century. This is a salutary historical parallel.I have long suspected that the US Deep State is being led by the nose by the malign Whitehall (the British Deep State): I call it the ‘master-blaster' syndrome. On August 6th, there was indirect confirmation of this in ex-British PM Boris Johnson's tweet about India. Let us remember he single-handedly ruined the chances of a peaceful resolution of the Ukraine War in 2022. Whitehall's mischief and meddling all over, if you read between the lines.Did I mention the British Special Force's views? Ah, Whitehall is getting a bit sloppy in its propaganda.Wait, so is India important (according to Whitehall) or unimportant (according to Trump)?Since I am very pro-American, I have a word of warning to Trump: you trust perfidious Albion at your peril. Their country is ruined, and they will not rest until they ruin yours too.I also wonder if there are British paw-prints in a recent and sudden spate of racist attacks on Indians in Ireland. A 6-year old girl was assaulted and kicked in the private parts. A nurse was gang-raped by a bunch of teenagers. Ireland has never been so racist against Indians (yes, I do remember the sad case of Savita Halappanavar, but that was religious bigotry more than racism). And I remember sudden spikes in anti-Indian attacks in Australia and Canada, both British vassals.There is no point in Indians whining about how the EU and America itself are buying more oil, palladium, rare earths, uranium etc. from Russia than India is. I am sorry to say this, but Western nations are known for hypocrisy. For example, exactly 80 years ago they dropped atomic bombs on Hiroshima and Nagasaki in Japan, but not on Germany or Italy. Why? The answer is uncomfortable. Lovely post-facto rationalization, isn't it?Remember the late lamented British East India Company that raped and pillaged India?Applying the three winning strategies to geo-economicsAs a professor of business strategy and innovation, I emphasize to my students that there are three broad ways of gaining an advantage over others: 1. Be the cost leader, 2. Be the most customer-intimate player, 3. Innovate. The US as a nation is patently not playing the cost leader; it does have some customer intimacy, but it is shrinking; its strength is in innovation.If you look at comparative advantage, the US at one time had strengths in all three of the above. Because it had the scale of a large market (and its most obvious competitors in Europe were decimated by world wars) America did enjoy an ability to be cost-competitive, especially as the dollar is the global default reserve currency. It demonstrated this by pushing through the Plaza Accords, forcing the Japanese yen to appreciate, destroying their cost advantage.In terms of customer intimacy, the US is losing its edge. Take cars for example: Americans practically invented them, and dominated the business, but they are in headlong retreat now because they simply don't make cars that people want outside the US: Japanese, Koreans, Germans and now Chinese do. Why were Ford and GM forced to leave the India market? Their “world cars” are no good in value-conscious India and other emerging markets.Innovation, yes, has been an American strength. Iconic Americans like Thomas Edison, Henry Ford, and Steve Jobs led the way in product and process innovation. US universities have produced idea after idea, and startups have ignited Silicon Valley. In fact Big Tech and aerospace/armaments are the biggest areas where the US leads these days.The armaments and aerospace tradeThat is pertinent because of two reasons: one is Trump's peevishness at India's purchase of weapons from Russia (even though that has come down from 70+% of imports to 36% according to SIPRI); two is the fact that there are significant services and intangible imports by India from the US, of for instance Big Tech services, even some routed through third countries like Ireland.Armaments and aerospace purchases from the US by India have gone up a lot: for example the Apache helicopters that arrived recently, the GE 404 engines ordered for India's indigenous fighter aircraft, Predator drones and P8-i Poseidon maritime surveillance aircraft. I suspect Trump is intent on pushing India to buy F-35s, the $110-million dollar 5th generation fighters.Unfortunately, the F-35 has a spotty track record. There were two crashes recently, one in Albuquerque in May, and the other on July 31 in Fresno, and that's $220 million dollars gone. Besides, the spectacle of a hapless British-owned F-35B sitting, forlorn, in the rain, in Trivandrum airport for weeks, lent itself to trolls, who made it the butt of jokes. I suspect India has firmly rebuffed Trump on this front, which has led to his focus on Russian arms.There might be other pushbacks too. Personally, I think India does need more P-8i submarine hunter-killer aircraft to patrol the Bay of Bengal, but India is exerting its buyer power. There are rumors of pauses in orders for Javelin and Stryker missiles as well.On the civilian aerospace front, I am astonished that all the media stories about Air India 171 and the suspicion that Boeing and/or General Electric are at fault have disappeared without a trace. Why? There had been the big narrative push to blame the poor pilots, and now that there is more than reasonable doubt that these US MNCs are to blame, there is a media blackout?Allegations about poor manufacturing practices by Boeing in North Charleston, South Carolina by whistleblowers have been damaging for the company's brand: this is where the 787 Dreamliners are put together. It would not be surprising if there is a slew of cancellations of orders for Boeing aircraft, with customers moving to Airbus. Let us note Air India and Indigo have placed some very large, multi-billion dollar orders with Boeing that may be in jeopardy.India as a consuming economy, and the services trade is hugely in the US' favorMany observers have pointed out the obvious fact that India is not an export-oriented economy, unlike, say, Japan or China. It is more of a consuming economy with a large, growing and increasingly less frugal population, and therefore it is a target for exporters rather than a competitor for exporting countries. As such, the impact of these US tariffs on India will be somewhat muted, and there are alternative destinations for India's exports, if need be.While Trump has focused on merchandise trade and India's modest surplus there, it is likely that there is a massive services trade, which is in the US' favor. All those Big Tech firms, such as Microsoft, Meta, Google and so on run a surplus in the US' favor, which may not be immediately evident because they route their sales through third countries, e.g. Ireland.These are the figures from the US Trade Representative, and quite frankly I don't believe them: there are a lot of invisible services being sold to India, and the value of Indian data is ignored.In addition to the financial implications, there are national security concerns. Take the case of Microsoft's cloud offering, Azure, which arbitrarily turned off services to Indian oil retailer Nayara on the flimsy grounds that the latter had substantial investment from Russia's Rosneft. This is an example of jurisdictional over-reach by US companies, which has dire consequences. India has been lax about controlling Big Tech, and this has to change.India is Meta's largest customer base. Whatsapp is used for practically everything. Which means that Meta has access to enormous amounts of Indian customer data, for which India is not even enforcing local storage. This is true of all other Big Tech (see OpenAI's Sam Altman below): they are playing fast and loose with Indian data, which is not in India's interest at all.Data is the new oil, says The Economist magazine. So how much should Meta, OpenAI et al be paying for Indian data? Meta is worth trillions of dollars, OpenAI half a trillion. How much of that can be attributed to Indian data?There is at least one example of how India too can play the digital game: UPI. Despite ham-handed efforts to now handicap UPI with a fee (thank you, brilliant government bureaucrats, yes, go ahead and kill the goose that lays the golden eggs), it has become a contender in a field that has long been dominated by the American duopoly of Visa and Mastercard. In other words, India can scale up and compete.It is unfortunate that India has not built up its own Big Tech behind a firewall as has been done behind the Great Firewall of China. But it is not too late. Is it possible for India-based cloud service providers to replace US Big Tech like Amazon Web Services and Microsoft Azure? Yes, there is at least one player in that market: Zoho.Second, what are the tariffs on Big Tech exports to India these days? What if India were to decide to impose a 50% tax on revenue generated in India through advertisement or through sales of services, mirroring the US's punitive taxes on Indian goods exports? Let me hasten to add that I am not suggesting this, it is merely a hypothetical argument.There could also be non-tariff barriers as China has implemented, but not India: data locality laws, forced use of local partners, data privacy laws like the EU's GDPR, anti-monopoly laws like the EU's Digital Markets Act, strict application of IPR laws like 3(k) that absolutely prohibits the patenting of software, and so on. India too can play legalistic games. This is a reason US agri-products do not pass muster: genetically modified seeds, and milk from cows fed with cattle feed from blood, offal and ground-up body parts.Similarly, in the ‘information' industry, India is likely to become the largest English-reading country in the world. I keep getting come-hither emails from the New York Times offering me $1 a month deals on their product: they want Indian customers. There are all these American media companies present in India, untrammelled by content controls or taxes. What if India were to give a choice to Bloomberg, Reuters, NYTimes, WaPo, NPR et al: 50% tax, or exit?This attack on peddlers of fake information and manufacturing consent I do suggest, and I have been suggesting for years. It would make no difference whatsoever to India if these media outlets were ejected, and they surely could cover India (well, basically what they do is to demean India) just as well from abroad. Out with them: good riddance to bad rubbish.What India needs to doI believe India needs to play the long game. It has to use its shatrubodha to realize that the US is not its enemy: in Chanakyan terms, the US is the Far Emperor. The enemy is China, or more precisely the Chinese Empire. Han China is just a rump on their south-eastern coast, but it is their conquered (and restive) colonies such as Tibet, Xinjiang, Manchuria and Inner Mongolia, that give them their current heft.But the historical trends are against China. It has in the past had stable governments for long periods, based on strong (and brutal) imperial power. Then comes the inevitable collapse, when the center falls apart, and there is absolute chaos. It is quite possible, given various trends, including demographic changes, that this may happen to China by 2050.On the other hand, (mostly thanks, I acknowledge, to China's manufacturing growth), the center of gravity of the world economy has been steadily shifting towards Asia. The momentum might swing towards India if China stumbles, but in any case the era of Atlantic dominance is probably gone for good. That was, of course, only a historical anomaly. Asia has always dominated: see Angus Maddison's magisterial history of the world economy, referred to below as well.I am reminded of the old story of the king berating his court poet for calling him “the new moon” and the emperor “the full moon”. The poet escaped being punished by pointing out that the new moon is waxing and the full moon is waning.This is the long game India has to keep in mind. Things are coming together for India to a great extent: in particular the demographic dividend, improved infrastructure, fiscal prudence, and the increasing centrality of the Indian Ocean as the locus of trade and commerce.India can attempt to gain competitive advantage in all three ways outlined above:* Cost-leadership. With a large market (assuming companies are willing to invest at scale), a low-cost labor force, and with a proven track-record of frugal innovation, India could well aim to be a cost-leader in selected areas of manufacturing. But this requires government intervention in loosening monetary policy and in reducing barriers to ease of doing business* Customer-intimacy. What works in highly value-conscious India could well work in other developing countries. For instance, the economic environment in ASEAN is largely similar to India's, and so Indian products should appeal to their residents; similarly with East Africa. Thus the Indian Ocean Rim with its huge (and in Africa's case, rapidly growing) population should be a natural fit for Indian products* Innovation. This is the hardest part, and it requires a new mindset in education and industry, to take risks and work at the bleeding edge of technology. In general, Indians have been content to replicate others' innovations at lower cost or do jugaad (which cannot scale up). To do real, disruptive innovation, first of all the services mindset should transition to a product mindset (sorry, Raghuram Rajan). Second, the quality of human capital must be improved. Third, there should be patient risk capital. Fourth, there should be entrepreneurs willing to try risky things. All of these are difficult, but doable.And what is the end point of this game? Leverage. The ability to compel others to buy from you.China has demonstrated this through its skill at being a cost-leader in industry after industry, often hollowing out entire nations through means both fair and foul. These means include far-sighted industrial policy including the acquisition of skills, technology, and raw materials, as well as hidden subsidies that support massive scaling, which ends up driving competing firms elsewhere out of business. India can learn a few lessons from them. One possible lesson is building capabilities, as David Teece of UC Berkeley suggested in 1997, that can span multiple products, sectors and even industries: the classic example is that of Nikon, whose optics strength helps it span industries such as photography, printing, and photolithography for chip manufacturing. Here is an interesting snapshot of China's capabilities today.2025 is, in a sense, a point of inflection for India just as the crisis in 1991 was. India had been content to plod along at the Nehruvian Rate of Growth of 2-3%, believing this was all it could achieve, as a ‘wounded civilization'. From that to a 6-7% growth rate is a leap, but it is not enough, nor is it testing the boundaries of what India can accomplish.1991 was the crisis that turned into an opportunity by accident. 2025 is a crisis that can be carefully and thoughtfully turned into an opportunity.The Idi Amin syndrome and the 1000 Talents program with AIThere is a key area where an American error may well be a windfall for India. This is based on the currently fashionable H1-B bashing which is really a race-bashing of Indians, and which has been taken up with gusto by certain MAGA folks. Once again, I suspect the baleful influence of Whitehall behind it, but whatever the reason, it looks like Indians are going to have a hard time settling down in the US.There are over a million Indians on H1-Bs, a large number of them software engineers, let us assume for convenience there are 250,000 of them. Given country caps of exactly 9800 a year, they have no realistic chance of getting a Green Card in the near future, and given the increasingly fraught nature of life there for brown people, they may leave the US, and possibly return to India..I call this the Idi Amin syndrome. In 1972, the dictator of Uganda went on a rampage against Indian-origin people in his country, and forcibly expelled 80,000 of them, because they were dominating the economy. There were unintended consequences: those who were ejected mostly went to the US and UK, and they have in many cases done well. But Uganda's economy virtually collapsed.That's a salutary experience. I am by no means saying that the US economy would collapse, but am pointing to the resilience of the Indians who were expelled. If, similarly, Trump forces a large number of Indians to return to India, that might well be a case of short-term pain and long-term gain: urvashi-shapam upakaram, as in the Malayalam phrase.Their return would be akin to what happened in China and Taiwan with their successful effort to attract their diaspora back. The Chinese program was called 1000 Talents, and they scoured the globe for academics and researchers of Chinese origin, and brought them back with attractive incentives and large budgets. They had a major role in energizing the Chinese economy.Similarly, Taiwan with Hsinchu University attracted high-quality talent, among which was the founder of TSMC, the globally dominant chip giant.And here is Trump offering to India on a platter at least 100,000 software engineers, especially at a time when generativeAI is decimating low-end jobs everywhere. They can work on some very compelling projects that could revolutionize Indian education, up-skilling and so on, and I am not at liberty to discuss them. Suffice to say that these could turbo-charge the Indian software industry and get it away from mundane, routine body-shopping type jobs.ConclusionThe Trump tariff tantrum is definitely a short-term problem for India, but it can be turned around, and turned into an opportunity, if only the country plays its cards right and focuses on building long-term comparative advantages and accepting the gift of a mis-step by Trump in geo-economics.In geo-politics, India and the US need each other to contain China, and so that part, being so obvious, will be taken care of more or less by default.Thus, overall, the old SWOT analysis: strengths, weaknesses, opportunities and threats. On balance, I am of the opinion that the threats contain in them the germs of opportunities. It is up to Indians to figure out how to take advantage of them. This is your game to win or lose, India!4150 words, 9 Aug 2025 This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit rajeevsrinivasan.substack.com/subscribe

AWS for Software Companies Podcast
Ep129: Taking Agentic AI Beyond the Prototype w Automation Anywhere

AWS for Software Companies Podcast

Play Episode Listen Later Aug 8, 2025 28:29


Industry leaders from Automation Anywhere and AWS discuss how modern customer data collection has evolved, and practical strategies for implementing enterprise automation at scale.Topics Include:Automation Anywhere and AWS experts discuss modern enterprise automation strategiesTraditional profiting strategies may not work with today's changing business modelsCustomer data collection methods have evolved across multiple platforms significantlyModern verification processes include automated validation systems and streamlined timelinesBackground check automation is increasingly handled by AI-powered models and systemsStanford's "Wonder Bread" research paper introduced revolutionary enterprise process observation technologyWonder Bread demonstrated AI systems watching and automatically learning hospital workflowsThe technology can author workflows by observing real enterprise processesEnterprise Process Management built around observed behaviors shows promising resultsVerification challenges exist since Wonder Bread research isn't widely publicized yetProcess observation technology could transform how enterprises handle workflow creationSalesforce Wizard Interface dominates many current automation implementations in enterprisesSalesforce Agent Codes offer alternative approaches to traditional automation methodsAWS platform selection involves careful consideration of enterprise integration needsDemo implementations showcase real-world timeline expectations and deployment maturity levelsCurrent automation solutions have reached significant scale across various industriesWorkflow automation differs fundamentally from true agentic intelligence systems capabilitiesAgentic AI demonstrates autonomous decision-making beyond simple rule-based automation processesUnderstanding this distinction helps organizations choose appropriate technology approaches effectivelySession concludes with clarity on modern automation landscape and implementation strategiesParticipants:Pratyush Garikapati – Director of Products, Automation AnywhereSreenath Gotur – Snr Generative AI Specialist, Amazon Web ServicesFurther Links:Automation Anywhere websiteAutomation Anywhere – AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

DailyCyber The Truth About Cyber Security with Brandon Krieger
Making Tech Work for SMBs with Steve Massaro | DailyCyber 271

DailyCyber The Truth About Cyber Security with Brandon Krieger

Play Episode Listen Later Aug 8, 2025 57:39


Making Tech Work for SMBs with Steve Massaro | DailyCyber 271 In this episode of DailyCyber, I'm joined by Steve Massaro, IT consultant, MSP owner, and strategic technologist with over 20 years of experience helping businesses succeed through managed services and secure IT leadership.Steve shares insight from his years building MSPs in Hawaii and Kansas—from scaling tech teams to keeping SMB clients safe in today's evolving threat landscape.

Order in the Court
To Fear or Not to Fear: The Fundamentals of AI and the Law

Order in the Court

Play Episode Listen Later Aug 7, 2025 46:02


On this episode, host Paul W. Grimm speaks with Professor Maura R. Grossman about the fundamentals of artificial intelligence and its growing influence on the legal system. They explore what AI is (and isn't), how machine learning and natural language processing work, and the differences between traditional automation and modern generative AI. In layman's terms, they discuss other key concepts, such as supervised and unsupervised learning, reinforcement training, and deepfakes, and other advances that have accelerated AI's development. Finally, they address a few potential risks of generative AI, including hallucinations, bias, and misuse in court, which sets the stage for a deeper conversation about legal implications on the next episode, "To Trust or Not to Trust: AI in Legal Practice." ABOUT THE HOSTJudge Paul W. Grimm (ret.) is the David F. Levi Professor of the Practice of Law and Director of the Bolch Judicial Institute at Duke Law School. From December 2012 until his retirement in December 2022, he served as a district judge of the United States District Court for the District of Maryland, with chambers in Greenbelt, Maryland. Click here to read his full bio.

AWS for Software Companies Podcast
Ep128: Co-Innovation in the Age of Agentic AI with Mark Relph of AWS

AWS for Software Companies Podcast

Play Episode Listen Later Aug 6, 2025 25:55


AWS's Mark Relph draws fascinating parallels between today's AI revolution and the 1900s agricultural mechanization that delivered 2,000% productivity gains, while exploring how agentic AI will fundamentally reshape every aspect of software business models.Topics Include:Mark Relph directs AWS's data and AI partner go-to-market strategy teamHis role focuses on making ISV partners a force multiplier for customer successPreviously ran go-to-market for Amazon Bedrock, AWS's fastest growing service everCurrent AI adoption pace exceeds even the early cloud computing boom yearsHistorical parallel: 1900s agricultural mechanization delivered 2,000% productivity gains and 95% resource reductionFirst commercial self-propelled farming equipment revolutionized entire economies and never looked back500 machines formed the "Harvest Brigade" during WWII, harvesting from Texas to CanadaMark has spoken to 600+ AWS customers about GenAI over two yearsOrganizations range from AI pioneers to those still "fending off pirates" internallyGenAI has become a phenomenal assistant within organizations for content and automationAWS's AI stack has three layers: infrastructure, Bedrock, and applicationsBottom layer provides complete control over training, inference, and custom applicationsMiddle layer Bedrock serves as the "operating system" for generative AI applicationsTop layer offers ready-to-use AI through Q assistants and productivity toolsAI systems are rapidly becoming more complex with multiple model chainsMany current "agents" are just really, really long prompts (Mark's hot take)Task-specific models are emerging as one size won't fit all use casesEvolution moves from human-driven AI to agent-assisted to fully autonomous agentsAgent readiness requires APIs that allow software to interact autonomouslyTraditional UIs become unnecessary when agents interface directly with systemsCore competencies shift when AI handles the actual "doing" of tasksSales and marketing must adapt to agents delivering outcomes autonomouslyGo-to-market strategies need complete rethinking for an agentic worldThe agentic age is upon us and AWS partners should shape the futureParticipants:Mark Relph – Director – Data & AI Partner Go-To-Market, Amazon Web ServicesSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Oracle University Podcast

In this episode, hosts Lois Houston and Nikita Abraham, together with Senior Cloud Engineer Nick Commisso, break down the basics of artificial intelligence (AI). They discuss the differences between Artificial General Intelligence (AGI) and Artificial Narrow Intelligence (ANI), and explore the concepts of machine learning, deep learning, and generative AI. Nick also shares examples of how AI is used in everyday life, from navigation apps to spam filters, and explains how AI can help businesses cut costs and boost revenue.   AI for You: https://mylearn.oracle.com/ou/course/ai-for-you/152601/252500   Oracle University Learning Community: https://education.oracle.com/ou-community   LinkedIn: https://www.linkedin.com/showcase/oracle-university/   X: https://x.com/Oracle_Edu   Special thanks to Arijit Ghosh, David Wright, Kris-Ann Nansen, Radhika Banka, and the OU Studio Team for helping us create this episode. ----------------------------------------------------------- Episode Transcript: 00:00 Welcome to the Oracle University Podcast, the first stop on your cloud journey. During this series of informative podcasts, we'll bring you foundational training on the most popular Oracle technologies. Let's get started! 00:25 Nikita: Hello and welcome to the Oracle University Podcast. I'm Nikita Abraham, Team Lead of Editorial Services with Oracle University, and with me is Lois Houston, Director of Innovation Programs. Lois: Hi everyone! Welcome to a new season of the podcast. I'm so excited about this one because we're going to dive into the world of artificial intelligence, speaking to many experts in the field. Nikita: If you've been listening to us for a while, you probably know we've covered AI from a bunch of different angles. But this time, we're dialing it all the way back to basics. We wanted to create something for the absolute beginner, so no jargon, no assumptions, just simple conversations that anyone can follow. 01:08 Lois: That's right, Niki. You don't need to have a technical background or prior experience with AI to get the most out of these episodes. In our upcoming conversations, we'll break down the basics of AI, explore how it's shaping the world around us, and understand its impact on your business. Nikita: The idea is to give you a practical understanding of AI that you can use in your work, especially if you're in sales, marketing, operations, HR, or even customer service.  01:37 Lois: Today, we'll talk about the basics of AI with Senior Cloud Engineer Nick Commisso. Hi Nick! Welcome back to the podcast. Can you tell us about human intelligence and how it relates to artificial intelligence? And within AI, I know we have Artificial General Intelligence, or AGI, and Artificial Narrow Intelligence, or ANI. What's the difference between the two? Nick: Human intelligence is the intellectual capability of humans that allow us to learn new skills through observation and mental digestion, to think through and understand abstract concepts and apply reasoning, to communicate using language and understand non-verbal cues, such as facial expressions, tone variation, body language. We can handle objections and situations in real time, even in a complex setting. We can plan for short and long-term situations or projects. And we can create music, art, or invent something new or have original ideas. If machines can replicate a wide range of human cognitive abilities, such as learning, reasoning, or problem solving, we call it artificial general intelligence.  Now, AGI is hypothetical for now, but when we apply AI to solve problems with specific, narrow objectives, we call it artificial narrow intelligence, or ANI. AGI is a hypothetical AI that thinks like a human. It represents the ultimate goal of artificial intelligence, which is a system capable of chatting, learning, and even arguing like us. If AGI existed, it would take the form like a robot doctor that accurately diagnoses and comforts patients, or an AI teacher that customizes lessons in real time based on each student's mood, pace, and learning style, or an AI therapist that comprehends complex emotions and provides empathetic, personalized support. ANI, on the other hand, focuses on doing one thing really well. It's designed to perform specific tasks by recognizing patterns and following rules, but it doesn't truly understand or think beyond its narrow scope. Think of ANI as a specialist. Your phone's face ID can recognize you instantly, but it can't carry on a conversation. Google Maps finds the best route, but it can't write you a poem. And spam filters catch junk mail, but it can't make you coffee. So, most of the AI you interact with today is ANI. It's smart, efficient, and practical, but limited to specific functions without general reasoning or creativity. 04:22 Nikita: Ok then what about Generative AI?  Nick: Generative AI is a type of AI that can produce content such as audio, text, code, video, and images. ChatGPT can write essays, but it can't fact check itself. DALL-E creates art, but it doesn't actually know if it's good. Or AI song covers can create deepfakes like Drake singing "Baby Shark."  04:47 Lois: Why should I care about AI? Why is it important? Nick: AI is already part of your everyday life, often working quietly in the background. ANI powers things like navigation apps, voice assistants, and spam filters. Generative AI helps create everything from custom playlists to smart writing tools. And while AGI isn't here yet, it's shaping ideas about what the future might look like. Now, AI is not just a buzzword, it's a tool that's changing how we live, work, and interact with the world. So, whether you're using it or learning about it or just curious, it's worth knowing what's behind the tech that's becoming part of everyday life.  05:32 Lois: Nick, whenever people talk about AI, they also throw around terms like machine learning and deep learning. What are they and how do they relate to AI? Nick: As we shared earlier, AI is the ability of machines to imitate human intelligence. And Machine Learning, or ML, is a subset of AI where the algorithms are used to learn from past data and predict outcomes on new data or to identify trends from the past. Deep Learning, or DL, is a subset of machine learning that uses neural networks to learn patterns from complex data and make predictions or classifications. And Generative AI, or GenAI, on the other hand, is a specific application of DL focused on creating new content, such as text, images, and audio, by learning the underlying structure of the training data.  06:24 Nikita: AI is often associated with key domains like language, speech, and vision, right? So, could you walk us through some of the specific tasks or applications within each of these areas? Nick: Language-related AI tasks can be text related or generative AI. Text-related AI tasks use text as input, and the output can vary depending on the task. Some examples include detecting language, extracting entities in a text, extracting key phrases, and so on.  06:54 Lois: Ok, I get you. That's like translating text, where you can use a text translation tool, type your text in the box, choose your source and target language, and then click Translate. That would be an example of a text-related AI task. What about generative AI language tasks? Nick: These are generative, which means the output text is generated by the model. Some examples are creating text, like stories or poems, summarizing texts, and answering questions, and so on. 07:25 Nikita: What about speech and vision? Nick: Speech-related AI tasks can be audio related or generative AI. Speech-related AI tasks use audio or speech as input, and the output can vary depending on the task. For example, speech to text conversion, speaker recognition, or voice conversion, and so on. Generative AI tasks are generative, i.e., the output audio is generated by the model (for example, music composition or speech synthesis). Vision-related AI tasks can be image related or generative AI. Image-related AI tasks use an image as the input, and the output depends on the task. Some examples are classifying images or identifying objects in an image. Facial recognition is one of the most popular image-related tasks that's often used for surveillance and tracking people in real time. It's used in a lot of different fields, like security and biometrics, law enforcement, entertainment, and social media. For generative AI tasks, the output image is generated by the model. For example, creating an image from a textual description or generating images of specific style or high resolution, and so on. It can create extremely realistic new images and videos by generating original 3D models of objects, such as machine, buildings, medications, people and landscapes, and so much more. 08:58 Lois: This is so fascinating. So, now we know what AI is capable of. But Nick, what is AI good at? Nick: AI frees you to focus on creativity and more challenging parts of your work. Now, AI isn't magic. It's just very good at certain tasks. It handles work that's repetitive, time consuming, or too complex for humans, like processing data or spotting patterns in large data sets.  AI can take over routine tasks that are essential but monotonous. Examples include entering data into spreadsheets, processing invoices, or even scheduling meetings, freeing up time for more meaningful work. AI can support professionals by extending their abilities. Now, this includes tools like AI-assisted coding for developers, real-time language translation for travelers or global teams, and advanced image analysis to help doctors interpret medical scans much more accurately. 10:00 Nikita: And what would you say is AI's sweet spot? Nick: That would be tasks that are both doable and valuable. A few examples of tasks that are feasible technically and have business value are things like predicting equipment failure. This saves downtime and the loss of business. Call center automation, like the routing of calls to the right person. This saves time and improves customer satisfaction. Document summarization and review. This helps save time for busy professionals. Or inspecting power lines. Now, this task is dangerous. By automating it, it protects human life and saves time. 10:48 Oracle University's Race to Certification 2025 is your ticket to free training and certification in today's hottest tech. Whether you're starting with Artificial Intelligence, Oracle Cloud Infrastructure, Multicloud, or Oracle Data Platform, this challenge covers it all! Learn more about your chance to win prizes and see your name on the Leaderboard by visiting education.oracle.com/race-to-certification-2025. That's education.oracle.com/race-to-certification-2025. 11:30 Nikita: Welcome back! Now one big way AI is helping businesses today is by cutting costs, right? Can you give us some examples of this?  Nick: Now, AI can contribute to cost reduction in several key areas. For instance, chatbots are capable of managing up to 50% of customer queries. This significantly reduces the need for manual support, thereby lowering operational costs. AI can streamline workflows, for example, reducing invoice processing time from 10 days to just 1 hour. This leads to substantial savings in both time and resources. In addition to cost savings, AI can also support revenue growth. One way is enabling personalization and upselling. Platforms like Netflix use AI-driven recommendation systems to influence user choices. This not only enhances the user experience, but it also increases the engagement and the subscription revenue. Or unlocking new revenue streams. AI technologies, such as generative video tools and virtual influencers, are creating entirely new avenues for advertising and branded content, expanding business opportunities in emerging markets. 12:50 Lois: Wow, saving money and boosting bottom lines. That's a real win! But Nick, how is AI able to do this?  Nick: Now, data is what teaches AI. Just like we learn from experience, so does AI. It learns from good examples, bad examples, and sometimes even the absence of examples. The quality and variety of data shape how smart, accurate, and useful AI becomes. Imagine teaching a kid to recognize animals using only pictures of squirrels that are labeled dogs. That would be very confusing at the dog park. AI works the exact same way, where bad data leads to bad decisions. With the right data, AI can be powerful and accurate. But with poor or biased data, it can become unreliable and even misleading.  AI amplifies whatever you feed it. So, give it gourmet data, not data junk food. AI is like a chef. It needs the right ingredients. It needs numbers for predictions, like will this product sell? It needs images for cool tricks like detecting tumors, and text for chatting, or generating excuses for why you'd be late. Variety keeps AI from being a one-trick pony. Examples of the types of data are numbers, or machine learning, for predicting things like the weather. Text or generative AI, where chatbots are used for writing emails or bad poetry. Images, or deep learning, can be used for identifying defective parts in an assembly line, or an audio data type to transcribe a dictation from a doctor to a text. 14:35 Lois: With so much data available, things can get pretty confusing, which is why we have the concept of labeled and unlabeled data. Can you help us understand what that is? Nick: Labeled data are like flashcards, where everything has an answer. Spam filters learned from emails that are already marked as junk, and X-rays are marked either normal or pneumonia. Let's say we're training AI to tell cats from dogs, and we show it a hundred labeled pictures. Cat, dog, cat, dog, etc. Over time, it learns, hmm fluffy and pointy ears? That's probably a cat. And then we test it with new pictures to verify. Unlabeled data is like a mystery box, where AI has to figure it out itself. Social media posts, or product reviews, have no labels. So, AI clusters them by similarity. AI finding trends in unlabeled data is like a kid sorting through LEGOs without instructions. No one tells them which blocks will go together.  15:36 Nikita: With all the data that's being used to train AI, I'm sure there are issues that can crop up too. What are some common problems, Nick? Nick: AI's performance depends heavily on the quality of its data. Poor or biased data leads to unreliable and unfair outcomes. Dirty data includes errors like typos, missing values, or duplicates. For example, an age record as 250, or NA, can confuse the AI. And a variety of data cleaning techniques are available, like missing data can be filled in, or duplicates can be removed. AI can inherit human prejudices if the data is unbalanced. For example, a hiring AI may favor one gender if the past three hires were mostly male. Ensuring diverse and representative data helps promote fairness. Good data is required to train better AI. Data could be messy, and needs to be processed before to train AI. 16:39 Nikita: Thank you, Nick, for sharing your expertise with us. To learn more about AI, go to mylearn.oracle.com and search for the AI for You course. As you complete the course, you'll find skill checks that you can attempt to solidify your learning.  Lois: In our next episode, we'll dive deep into fundamental AI concepts and terminologies. Until then, this is Lois Houston… Nikita: And Nikita Abraham signing off! 17:05 That's all for this episode of the Oracle University Podcast. If you enjoyed listening, please click Subscribe to get all the latest episodes. We'd also love it if you would take a moment to rate and review us on your podcast app. See you again on the next episode of the Oracle University Podcast.

On Record PR
AI Just Changed Search Forever: Here's What Law Firms Need to Know

On Record PR

Play Episode Listen Later Aug 4, 2025 15:00


Building on last week's DigiMarCon 2025 recap, Leslie Richards and Isabelle Horan unpack the evolution from traditional Search Engine Optimization (SEO) to Generative Engine Optimization (GEO) and what that means for law firms today.

AWS for Software Companies Podcast
Ep127: Enabling AI Acceleration at Scale - How Celonis Leverages Amazon Bedrock

AWS for Software Companies Podcast

Play Episode Listen Later Aug 4, 2025 50:13


Industry leaders from Celonis and AWS explain why 2025 marks the inflection point for agentic AI and how early adopters are gaining significant competitive advantages in efficiency and innovation.Topics Include:AWS's Cristen Hughes and Celonis's Jeff Naughton discuss AI agent transformationAndy Jassy declares AI agents will fundamentally change how we workThree key trends make AI agents practical: smarter models, longer tasks, cheaper costsAI now beats humans on complex benchmarks for the first time everClaude 3.7 cracked graduate-level reasoning where humans previously dominated completelyAI evolved from brief interactions to managing sustained multi-step complex workflowsProcessing costs plummeted 99.7% making enterprise-grade AI economically viable at scaleWe're transitioning from 2023's adaptation era to 2025's human-AI collaboration eraBy 2028, AI will suggest actions to humans rather than vice versaAgents are autonomous software that plan, act, and reason independently with minimal interventionAgent workflow: receive human request, create plan, execute actions, review, adjust, deliverFour agent components: brain (LLM), memory (context), actions (tools), persona (role definition)AWS offers three building approaches: ready-made solutions, managed platform, DIY developmentKey enterprise applications: software development acceleration, customer care automation, knowledge work optimizationManual processes like accounts payable offer huge transformation opportunities through intelligent automationDeep process analysis is critical before deploying agents for maximum effectivenessCelonis pioneered process mining to help enterprises understand their actual workflow realitiesCompanies are collections of interacting processes that agents need proper context to navigateProcess intelligence provides agents with placement guidance, data feeds, monitoring, and workflow directionCelonis-AWS partnership demonstrates order management agents that automatically handle at-risk situationsParticipants:Jeff Naughton – SVP and Fellow, CelonisCristen Hughes – Solutions Architecture Leader, ISV, North America, Amazon Web ServicesFurther Links:Celonis WebsiteCelonis on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep126: Using AWS to Transform Customer Interactions with Glia

AWS for Software Companies Podcast

Play Episode Listen Later Aug 1, 2025 13:53


Justin DiPietro, Co-Founder & Chief Strategy Officer of Glia, shares how they are leveraging AI to enhance the customer experience in the highly regulated world of financial institutions.Topics Include:Glia provides voice, digital, and AI services for customer-facing and internal operationsBuilt on "channel-less architecture" unlike traditional contact centers that added channels sequentiallyOne interaction can move seamlessly between channels (voice, chat, SMS, social)AI applies across all channels simultaneously rather than per individual channel700 customers, primarily banks and credit unions, 370 employees, headquartered in New YorkTargets 3,500 banks and credit unions across the United States marketFocuses exclusively on financial services and other regulated industriesAI for regulated industries requires different approach than non-regulated businessesTraditional contact centers had trade-off between cost and quality of serviceAI enables higher quality while simultaneously decreasing costs for contact centersNumber one reason people call banks: "What's my balance?" (20% of calls)Financial services require 100% accuracy, not 99.999% due to trust requirementsUses AWS exclusively for security, reliability, and future-oriented technology accessReal-time system requires triple-hot redundancy; seconds matter for live callsWorks with Bedrock team; customers certify Bedrock rather than individual featuresShowed examples of competitors' AI giving illegal million-dollar loans at 0%"Responsible AI" separates probabilistic understanding from deterministic responses to customersUses three model types: client models, network models, and protective modelsTraditional NLP had 50% accuracy; their LLM approach achieves 100% understandingPolicy is "use Nova unless" they can't, primarily for speed benefitsParticipants:Justin DiPietro – Co-Founder & Chief Strategy Officer, GliaFurther Links:Glia WebsiteGlia AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep125: Bridging the gap between requirements and budget - Better data while still controlling costs

AWS for Software Companies Podcast

Play Episode Listen Later Jul 30, 2025 25:39


Ed Bailey, Field CISO at Cribl, shares how Cribl and AWS are helping customers rethink their data strategy by making it easier to modernize, reduce complexity, and unlock long-term flexibility.Topics Include:Ed Bailey introduces topic: bridging gap between security data requirements and budgetCompanies face mismatch: 10TB data needs vs 5TB licensing budget constraintsData volumes growing exponentially while budgets remain relatively flat year-over-yearIT security data differs from BI: enormous volume, variety, complexityMany companies discover 600+ data sources during SIEM migration projects50% of SIEM data remains un-accessed within 90 days of ingestionComplex data collection architectures break frequently and require excessive maintenanceTeams spend 80% time collecting data, only 20% analyzing for valueData collection and storage are costs; analytics and insights provide business valuePoor data quality creates operational chaos requiring dozens of browser tabsSOC analysts struggle with context switching across multiple disconnected systemsTraditional vendor approach: "give us all data, we'll solve problems" is outdatedData modernization requires sharing information widely across organizational business unitsData maturity model progression: patchwork → efficiency → optimization → innovationData tiering strategy: route expensive SIEM data vs cheaper data lake storageSIEM costs ~$1/GB while data lakes cost ~$0.15-0.20/GB for storageCompliance retention data should go to object storage at penny fractionsDecouple data retention from vendor tools to enable migration flexibilityCribl platform offers integrated solutions: Stream, Search, Lake, Edge componentsCustomer success: Siemens reduced 5TB to 500GB while maintaining security effectivenessParticipants:Edward Bailey – Field CISO, CriblFurther Links:Cribl WebsiteCribl on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

The Wall Street Resource
Soluna Holdings, Inc. (SLNH) John Belizaire, CEO

The Wall Street Resource

Play Episode Listen Later Jul 29, 2025 30:25


The Company designs, develops, and operatesdigital infrastructure that transforms surplus renewable energy into globalcomputing resources. Soluna's pioneering data centers are strategicallyco-located with wind, solar, or hydroelectric power plants to supporthigh-performance computing applications, including Bitcoin Mining, GenerativeAI, and other compute-intensive applications. Soluna's proprietary softwareMaestroOS(™) helps energize a greener grid while delivering cost-effective andsustainable computing solutions and superior returns. A leading developer of green data centers that convert excess renewableenergy into global computing resources. Soluna builds modular, scalable datacenters for computing intensive, batchable applications such as Bitcoin mining,AI, and machine learning. Soluna provides a cost-effective alternative tobattery storage or transmission lines. Up to 30% of the power of renewableenergy projects can go to waste. Soluna's data centers enable clean electricityasset owners to ‘Sell. Every. Megawatt.'

On Record PR
How AI is Reshaping Digital Marketing and Law Firm Strategy

On Record PR

Play Episode Listen Later Jul 28, 2025 15:14


Leslie Richards and Isabelle Horan share key takeaways from the 2025 DigiMarCon Conference, including how law firms can move beyond basic content generation and instead use generative AI tools for advanced strategic applications.

AWS for Software Companies Podcast
Ep124: Powering Enterprise AI - How Our AI Journey Evolved featuring Jamf

AWS for Software Companies Podcast

Play Episode Listen Later Jul 28, 2025 28:03


Sam Johnson, Chief Customer Officer of Jamf, discusses the implementation of AI built on Amazon Bedrock that is a gamechanger in helping Jamf's 76,000+ customers scale their device management operations.Topics Include:Sam Johnson introduces himself as Chief Customer Officer from Jamf companyJamf's 23-year mission: help organizations succeed with Apple device managementCompany manages 33+ million devices for 76,000+ customers worldwide from MinneapolisJamf has used AI since 2018 for security threat detectionReleased first customer-facing generative AI Assistant just last year in 2024Presentation covers why, how they built it, use cases, and future plansJamf serves horizontal market from small business to Fortune 500 companiesChallenge: balance powerful platform capabilities with ease of use and adoptionAI could help get best of both worlds - power and simplicityAI also increases security posture and scales user capabilities significantlyCustomers already using ChatGPT/Claude but wanted AI embedded in productBuilt into product to reduce "doorway effect" of switching digital environmentsCreated small cross-functional team to survey land and build initial trailRest of engineering organization came behind to build the production highwayTeam needed governance layer with input from security, legal, other departmentsEvaluated multiple providers but ultimately chose Amazon Bedrock for three reasonsAWS team support, large community, and integration with existing infrastructureUses Lambda, DynamoDB, CloudWatch to support the Bedrock AI implementationAI development required longer training/validation phase than typical product featuresReleased "AI Assistant" with three skills: Reference, Explain, and Search capabilitiesParticipants:Sam Johnson – Chief Customer Officer, JamfFurther Links:Jamf.comJamf on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep123: Signal from the Noise - How SecurityScorecard leverages AI to Power Global Threat Detection

AWS for Software Companies Podcast

Play Episode Listen Later Jul 25, 2025 17:22


Mark Stevens, SVP, Channels and Alliances, discusses how SecurityScorecard's strategic partnership with AWS enables them to scale their security solutions through cloud infrastructure, marketplace integration, and co-sell programsTopics Include:SecurityScorecard founded 10 years ago to understand third-party vendor security postureCompany has grown to 3,000 enterprise customers and 200+ partners globallyEvolved from ratings to "supply chain detection and response" over last yearSupply chain threats have doubled, creating extended attack surfaces for companiesMany organizations don't know their vendor count or vulnerabilities within supply chainsSecurityScorecard provides visibility into attack surfaces and management tools for controlGenerative AI is central to their ecosystem, leveraging AWS Bedrock extensivelyThey scan the entire internet every two days at massive scaleHave scored 12 million companies with security scorecards to dateAll workloads run on AWS cloud infrastructure as their primary platformAWS partnership provides necessary scale for managing hundreds of thousands of vendorsCase study: Identified vendor misconfigurations that could shut down 1,000 locationsOwn massive 10-year data lake with tens of millions of companiesNew managed service combines AI automation with human analysts for supportLarge organizations cannot fully automate supply chain security management yetQuality threat intelligence data now valuable to SOC teams, not just riskThird-party risk management and SOC teams are slowly converging for better securityAWS marketplace integration provides frictionless customer experience and larger dealsCo-sell programs with AWS enterprise sales teams create effective flywheel motionFuture expansion includes identity management, response actions, and internal signal managementParticipants:Mark Stevens – SVP, Channels and Alliances, SecurityScorecardFurther Links:SecurityScorecard.ioSecurityScorecard AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Legal Marketing Minutes with Nancy Myrland
074: Lawyers Beware - There Could Be Serious Ethics Issues With The New AI Browsers

Legal Marketing Minutes with Nancy Myrland

Play Episode Listen Later Jul 24, 2025 9:43


AI just took a major step forward, and lawyers need to be paying attention. In this episode, I break down the launch of new AI browsers from ChatGPT and Perplexity. These tools are designed to act on your behalf, reading files, scanning emails, analyzing data, and even taking action while running quietly in the background. If that sounds efficient, it also raises serious red flags for anyone who handles sensitive or confidential information. We'll cover: • What agentic AI browsers really do • Why this technology goes beyond typical search • What OpenAI's Sam Altman is saying about the risks • How ABA Model Rule 1.6 applies here • What actions firms should take right now to protect confidentiality This is not about avoiding new technology. It is about using it wisely and protecting your clients in the process.

Thoughts on the Market
Will the Entertainment Business Stay Human?

Thoughts on the Market

Play Episode Listen Later Jul 23, 2025 5:15


Our U.S. Media & Entertainment Analyst Benjamin Swinburne discusses how GenAI is transforming content creation, distribution and also raising some serious ethical questions. Read more insights from Morgan Stanley.----- Transcript -----Welcome to Thoughts on the Market. I'm Ben Swinburne, Morgan Stanley's U.S. Media and Entertainment Analyst. Today – GenAI is poised to shake up the entertainment business. It's Wednesday, July 23, at 10am in New York.It's never been easier to create art for anyone – with a little help from GenerativeAI. You can transform photos of yourself or loved ones in the style of a popular Japanese movie studio or any era of visual art to your liking. You can create a short movie by simply typing in a few prompts. Even I can speak to youin several different languages. I can ask about the weather:Hvordan er været i dag?Wie ist das wetter heute?आज मौसम कैसा है? In the media and entertainment industry, GenAI is expected to bring about a seismic shift in how content is made and consumed. A recent production used AI to de-age actors and recreate the likeness of a deceased performer—cutting what used to take hundreds of VFX artists a year to just a few months with a small team. There are many other examples of how GenAI is revolutionizing how stories are told, from scriptwriting and editing to visual effects and dubbing. In music, GenAI is helping music labels identify emerging talent and generate new compositions. GenAI can even create songs using the voices of long-gone artists – potentially extending revenue far beyond an artist's lifetime. GenAI-driven tools have the potential to reduce TV and film production costs by 10–30 percent, with animation and post-production among the biggest savings opportunities. GenAI could also transform how content reaches audiences. Recommendation engines can become even more predictive, using behavioral data to serve up exactly what listeners want—sometimes before we know what we want. And there's more studios can achieve in post production. GenAI can already dub content in multiple languages, even syncing mouth movements to match the new dialogue. This makes global distribution faster, cheaper, and more culturally relevant. With better engagement comes better monetization. Platforms will use GenAI to introduce new pricing tiers, targeted advertising, and personalized superfan content that taps into niche audiences willing to pay more. But all this innovation brings up profound ethical concerns. First, there's the issue of consent and copyright. Can GenAI tools legally use an actor's name, likeness or voice? Then there's the question of authorship. If an AI writes a script or composes a song, who owns the rights? The creator or the GenAI model? Labor unions are understandably worried. In 2023, AI was a major sticking point in negotiations between Hollywood studios and writers' and actors' guilds. The fear? That AI could replace human jobs or devalue creative work. There are also legal battles. Multiple lawsuits are underway over whether AI models trained on copyrighted material without permission violate intellectual property laws. The outcomes of these cases could reshape the entire industry. But here's a big question no one can ignore: Will audiences care if content is AI-generated? Some consumers are fascinated by AI-created music or visuals, while others crave the emotional depth and authenticity that comes from human storytelling. Made-by-humans could become a premium label in itself. Now, despite GenAI's rapid rise, not every corner of entertainment is vulnerable. Live sports, concerts, and theater remain largely insulated from AI disruption. These experiences thrive on real-time emotion, unpredictability, and human connection—things AI can't replicate. In an AI-saturated world, the value of live events and sports rights will rise, favoring owners of sports rights and live platforms. So where do we go from here? By and large, we're entering an era where storytelling is no longer limited by budget or geography. GenAI is lowering the barriers to entry, expanding the creative class, and reshaping the economics of media. The winners in this new landscape will likely be companies that can scale—platforms with massive user bases, deep data pools, and the engineering talent to integrate GenAI seamlessly. But there's also room for agile newcomers who can innovate faster than the incumbents and disrupt the disrupters. No doubt, as the tools get better, the questions get harder. And that's where the real story begins. Thanks for listening. If you enjoy the show, please leave us a review wherever you listen and share Thoughts on the Market with a friend or colleague today.

AWS for Software Companies Podcast
Ep122: Securing the Software Supply Chain - How Sonatype Protects Developers in the Age of AI

AWS for Software Companies Podcast

Play Episode Listen Later Jul 23, 2025 19:54


Chief Product Development Officer Mitchell Johnson discusses how Sonatype protects enterprise developers from malicious open source components while keeping them productive through AI.Topics Include:Sonatype provides software supply chain solutions for enterprises using open source componentsThey serve large enterprises, government agencies, and critical infrastructure providers globallyMain challenge: keeping developers productive while maintaining secure software supply chainsCybercrime and supply chain attacks are massive, growing industries threatening developersAI adoption is happening faster than expected, profoundly changing development workflowsBad actors evolved from waiting for vulnerabilities to creating malicious componentsMalicious open source components specifically target developer and DevOps toolchainsSonatype's security research team uses AI/ML to analyze every open source componentThey can predict and block malicious components before entering customer environmentsAWS partnership helps Sonatype meet customers where they want to do businessPartnership focuses on go-to-market alignment, not just technical integrationAWS sales teams should be treated as extensions of your own sales organizationUnderstanding AWS sales structure and incentives is crucial for successful partnershipsAI development is following same pattern as open source adoption twenty years ago"Shadow AI" parallels the earlier "shadow IT" trend with open source softwareAI speeds up code generation but security review processes haven't kept paceDevelopers need a "Hippocratic Oath" - taking responsibility for AI-generated code outputWithin 24 months, professionals not skilled in AI will struggle to stay relevantSonatype's culture encourages curiosity, experimentation, and accepts failure as part of innovationTheir core mission: help developers focus on innovation, not security choresParticipants:Mitchell Johnson – Chief Product Development Officer, SonatypeFurther Links:Sonatype WebsiteSonatype on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep121: Ethical Hackers and AI Agents: The Future of Vulnerability Management with HackerOne

AWS for Software Companies Podcast

Play Episode Listen Later Jul 21, 2025 19:54


Founder and CTO Alex Rice discusses how HackerOne uses generative AI to automate security workflows and prioritizing accuracy over efficiency to achieve end-to-end outcomes.Topics Include:HackerOne uses ethical hackers and AI to find vulnerabilities before criminalsWhite hat hackers stress test systems to identify security weaknesses proactivelyGenerative AI plays a huge role in HackerOne's security operationsSecurity teams struggle with constant toil of finding and fixing vulnerabilitiesAI helps minimize toil through natural language interfaces and automationBoth good and bad actors have access to generative AI toolsSuccess requires measuring individual task inputs and outputs, not just aggregatesBreaking down workflows into granular tasks reveals measurable AI improvementsHackerOne deployed "Hive," their AI security agent to reduce customer toilInitial focus was on tasks where AI clearly outperformed humansStarted with low-hanging fruit before tackling more complex strategic workflowsAccuracy is the primary success metric, not just efficiency or speedSecurity requires precision; wrong fixes create bigger problems than inefficiencyCustomer acceptance and reduced time to remediation are north star metricsHumans remain the source of truth for validation and feedback loopsBreak down human jobs into granular AI tasks using systems thinkingBuild specific agents for individual tasks rather than entire job rolesKeep humans accountable for end-to-end outcomes to maintain customer trustAWS Bedrock chosen for security, confidentiality, and data separation requirementsMoving from efficiency improvements to entirely new AI-enabled capabilitiesParticipants:Alex Rice – Founder & CTO/CISO, HackerOneFurther Links:HackerOne WebsiteHackerOne on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

The Human Risk Podcast
Zsike Peter on Thinkbait

The Human Risk Podcast

Play Episode Listen Later Jul 19, 2025 69:45


What if the real risk of AI isn't job loss but brain atrophy?Episode SummaryIf you've spent any time on social media recently, you'll be familiar with the flood of low-quality AI-generated sludge. And on this episode, I'm speaking to someone who is leading a one-woman campaign against it and in favour of human-generated content. Her name is Zsike Peter and she's the fiercely human founder of an agency called Vampire Digital; you'll hear why its called that on the show.  Zsike is also the author of a new book called Thinkbait: The Definitive Guide to Reclaiming Human Creativity in the Age of AI which seeks to highlight and combat the prevalence of AI-generated low-quality content.Her mission is a passionate defence of human creativity in an age where generative AI threatens to dull our minds and voices. Its a rallying cry for intentional, thoughtful use that keeps our agency intact. In a fascinating discussion, we explore her extraordinary backstory, from growing up in communist Transylvania to being arrested after escaping a toxic UK host family that hired her as an au pair. And you'll hear the remarkable story about how she went undercover in a brothel to win a journalism scholarship. These stories aren't just great anecdotes, they reveal a mindset of relentless curiosity, courage, and independence that informs her work today.What makes Zsike's objection to AI so compelling is that initially she embraced it.  But after having tried it out, she flipped from embracing generative AI to warning against its effects on our thinking. You'll hear her talk about the concept of Thinkbait as an alternative to clickbait; content that stimulates rather than stupefies. Along the way, we unpack how language creates culture, why writing is thinking, and what it means to preserve our humanness in a machine-saturated world.And much, much more.Guest Biography: Zsike Peter Zsike is the founder of Vampire Digital — a content agency with a “fiercely human heart,” known for producing sharp, human-written copy in a world drowning in AI sludge. She is also the author of Thinkbait: The Definitive Guide to Reclaiming Human Creativity in the Age of AI — a book that serves as both a practical framework and a philosophical manifesto. Her background in journalism, corporate communications, and digital marketing now powers a mission to help people reclaim their voices (and their thinking) in a world increasingly seduced by generative AI.LinksThinkbait  - https://thinkbait.co.uk/Vampire Digital, Zsike's agency - https://www.vampiredigital.biz/Zsike on LinkedIn - https://www.linkedin.com/in/zsike-peter/AI-Generated Timestamped Summary(yes, I know, ironic, given the subject!)[00:00:00] Introduction [00:02:00] Zsike's childhood in communist Transylvania and family escape story[00:13:00] Going undercover in a brothel to win a journalism competition[00:19:00] Her arrest and start in the UK after fleeing abuse[00:24:00] Building a career in communications and founding Vampire Digital[00:28:00] Why she chose the vampire brand and what it represents[00:31:00] How her agency captures authentic voice in client content[00:33:00] Her shift from embracing to warning against generative AI[00:36:00] The dangers of outsourcing thinking and writing to machines[00:41:00] Why individuality and voice matter in a world of sameness[00:44:00] Thinkbait as a framework, manifesto, and act of defiance[00:48:00] The bedtime story moment that triggered a rethink on AI[00:53:00] The rise of fake authority and automated engagement online[00:57:00] Language loss, writing in a third language, and cultural identity[01:03:00] How hardship shaped her creative drive and ethical stance[01:07:00] Final reflections

AWS for Software Companies Podcast
Ep120: Asana and Amazon Q - Co-Innovating with AWS Generative AI Services

AWS for Software Companies Podcast

Play Episode Listen Later Jul 17, 2025 27:37


Spencer Herrick, Principal AI Product Manager of Asana and Oliver Myers of AWS demonstrate how their integration allows Asana's AI workflows to access enterprise data from Amazon Q Business, enabling seamless cross-application automation and insights.Topics Include:Oliver Myers leads Amazon Q Business go-to-market, Spencer Herrick manages Asana AI products.Session focuses on end user productivity challenges with generative AI technology implementations.End users face technology overload with doubled workplace application usage over five years.Data silos prevent getting maximum value from generative AI across fragmented enterprise systems.Workers spend 53% of time on "work about work" instead of strategic contributions.Ideal experience needs single pane of glass with cross-application insights and actions.Amazon Q Business launched as managed service with 40+ enterprise data connectors.Connectors maintain end-user permissions from source systems for enterprise security compliance.QIndex feature enables ISVs to access Q Business data via API calls.End users get answers enriched with multiple data sources without switching applications.Asana's work graph connects all tasks, projects, and portfolios to company goals.Phase 1 AI focused on narrow solutions like smart status updates.Phase 2 aimed for AI teammate capabilities requiring extensive contextual knowledge.AI Studio launched as no-code workflow automation builder within Asana platform.Q integration allows AI Studio to access cross-application context beyond Asana boundaries.SmartChat enhanced with Q can answer "what should I work on today?" holistically.Users returning from PTO can quickly understand goal risks across data sources.AI Studio workflows automate feature request processing across Asana, Drive, Slack, email.Partnership eliminates silos while maintaining enterprise security and permission controls.Integration creates connected ecosystem enabling true cross-application AI automation and insights.Participants:Spencer Herrick - Principal AI Product Manager, AsanaOliver Myers - Worldwide Head of Business Development, Amazon Web ServicesFurther Links:Asana.comAsana on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep119: Process Intelligence in the Age of AI – A New Era of Business Automation with Celonis

AWS for Software Companies Podcast

Play Episode Listen Later Jul 16, 2025 24:31


Chief Product Officer Dan Brown explains how Celonis creates digital twins of business processes to power AI agents that automate operational improvements.Topics Include:Dan Brown introduces Celonis as the thought leader in process mining for over a decade.Celonis serves largest global companies across all industries seeking operational improvements.Companies have process diagrams but actual operations differ significantly from documentation.Celonis creates digital twins of business processes by analyzing system data flows.Process intelligence reveals how work actually happens versus how companies think it happens.Platform enables process normalization, improvement assessment, and automated corrective actions.Celonis vision: making processes work better for people, companies, and the planet.Process intelligence provides visibility into current operations and improvement strategies.Great AI requires great data, but most companies only have static views.Process intelligence graph shows real-time flow of orders, invoices, and opportunities.Agentic AI requires four capabilities: sensing, planning, executing, and governing operations.Process intelligence enables real-time detection of conformance problems and deviations.AWS partnership leverages Bedrock for agentic AI and infrastructure for data processing.Data ingestion, organization, and enrichment are core to process intelligence value.AI agents now handle process deviations with increasing autonomy and sophistication.Heavy equipment manufacturer uses agents to coordinate with third-party vendors automatically.Agents text and email vendors to confirm delivery dates, reducing manual work.Implementation challenges include data quality, conservative adoption, and governance concerns.Companies should start with achievable use cases and expand gradually across domains.Future involves enterprise-wide process visibility powering automated applications and continuous improvement.Participants:Dan Brown – Chief Product Officer, CelonisFurther Links:Celonis WebsiteCelonis on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep118: Revolutionizing Customer Experience through Generative AI with Automation Anywhere, Qlik and Vectra.ai

AWS for Software Companies Podcast

Play Episode Listen Later Jul 14, 2025 46:56


AWS partners Automation Anywhere, Qlik, and Vectra.ai discuss revolutionizing customer experience through generative AI, sharing real-world implementations in automation, analytics, and cybersecurity applications. Topics Include:AWS Technology Partnerships panel on agentic AI implementationThree AWS partners share real-world AI deployment experiencesAutomation Anywhere automates end-to-end business processes with agentsVectra.ai uses autonomous agents for cybersecurity threat detectionQlik applies generative AI across their data platform portfolioCustomer service automation handles L1 support requests efficientlyUtility company processes 144,000 complaints annually using agentsRegulatory compliance improved through faster complaint resolution workflowsCybersecurity agents reduce threat detection time by 50-60%Triage, correlation, and prioritization handled by autonomous agentsSignal fatigue reduced through intelligent alert filtering systemsNatural language queries enable faster business decision makingSales AI agent provides competitive information during callsAWS Marketplace reduced 7,000 weekly tickets to zero2023 was proof-of-concept year, 2024 focuses production deploymentAWS Bedrock integration seamless with existing data repositoriesModel optionality crucial for different use case requirementsAgility most important capability in generative AI journeyCode abandonment becomes acceptable due to rapid innovationMaximum team size of 10 people maintains development agilityTargeted solutions outperform broad platform capabilities in adoptionImplementation expertise becomes bottleneck for customer scaling effortsNatural language interaction patterns completely shifted user behaviorKeywords searches replaced by conversational query approachesResponsible AI committees review decisions and establish principlesSecurity considerations balance speed with responsible deployment practicesBad actors adopt generative AI faster than defendersExplainability requirements slow feature rollout but ensure auditabilityMulti-modal deployments use different models for specific casesFuture trends include AI-powered business process outsourcingParticipants:Peter White – SVP, Emerging Products, Automation AnywhereRyan Welsh – Field CTO - Generative AI, QlikJohn Skinner – Vice President Corporate/Business Development, Vectra.aiChris Grusz – Managing Director for Technology Partnerships, AWSFurther Links:Automation Anywhere in AWS MarketplaceQlik in AWS MarketplaceVectra.ai in AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep117: Breaking Down Silos: Trellix's AI-Driven Security Operations

AWS for Software Companies Podcast

Play Episode Listen Later Jul 10, 2025 16:43


Zak Krider, Trellix's Director of Strategy and AI, shares how Trellix has successfully integrated generative AI into their security operations and democratized access to AI models across the organization.Topics Include:Trellix formed from McAfee Enterprise and FireEye mergerProvides full security stack visibility in single platformServes SMBs to Fortune 500 and government customersUsed machine learning for two decades with 30 modelsRecently pivoted to generative AI with Wwise platformAI finds critical events among thousands daily alertsIncorporates threat hunting knowledge into AI prompt structuresAWS Bedrock central to AI strategy for model flexibilityFormed small tiger team to investigate generative AIAnthropic Claude provided breakthrough "aha moments" for capabilitiesAdopted "fail fast, learn fast" innovation culture approachEnabled employee access to models through Bedrock APIConducted innovation jam sessions with VC-style pitchesAI decoded Base64 without prompting, identified benign activityJunior analysts elevated to level two with AICommon misconception: models train on customer data falselyEarly challenge: providing too much data overwhelmed modelsSmaller models hallucinated more with plausible-sounding responsesDesign partner programs help prioritize product developmentDemocratize AI access beyond just technical teamsTest multiple models for specific use casesLarge models work better than small ones initiallyPrompt engineering crucial for effective model communicationModel Context Protocol will gain traction next yearBackend data security remains largely unsolved challengeFederal customers require on-premises, air-gapped AI solutionsParticipants:Zak Krider – Director of AI and Innovation, TrellixFurther Links:Website: https://www.trellix.comTrellix on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep116: Building the AI Economy - Inside NVIDIA's 25,000-Strong Startup Ecosystem

AWS for Software Companies Podcast

Play Episode Listen Later Jul 9, 2025 12:52


NVIDIA's Global Head of Partnerships & Cloud for Startups, Jen Hoskins, details their collaboration with AWS to support over 25,000 startups through their Inception program.Topics Include:AI transformation happening across all industries and verticalsNVIDIA evolved from GPU company to full-stack AI solutionsAccelerated computing requires complete stack re-engineering from chip upTraditional CPU scaling has reached its fundamental performance limitsNVIDIA-AWS partnership spans over 13 years of co-developmentDGX Cloud integrates seamlessly with AWS SageMaker and BedrockOver 26 NVIDIA solutions available in AWS MarketplaceNVIDIA AI Enterprise accelerates data science and deployment pipelinesNIM microservices streamline AI model development like Docker containersCodeway gaming startup saved 48% on compute costs using NVIDIAEternal improved marketing ROI by 30X with generative AIQuoto achieved 10X content length and 3X throughput improvementNOATech biotech scaled cancer research with small team efficientlyNVIDIA Inception program supports over 25,000 startups globallyProgram covers 100+ countries across all verticals and stagesStartups get AWS credits up to $100,000 through ActivateDeveloper program offers free access to hundreds of SDKsThree program pillars: Innovate, Build, and Grow stagesVC Alliance connects startups with over 1,000 investorsVenture Capital Connect directly links startups to funding opportunitiesParticipants:Jen Hoskins – Startups, Global Head of Cloud, Partnerships & Go to Market, NVIDIAFurther Links:Website: https://www.nvidia.comNVIDIA Inception ProgramNVIDIA on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

AWS for Software Companies Podcast
Ep115: Put AI to Work Supercharging Enterprise Intelligence with Glean + AWS

AWS for Software Companies Podcast

Play Episode Listen Later Jul 7, 2025 16:59


Matt “Kix” Kixmoeller, Chief Marketing Officer of Glean, shares how Glean partners with AWS to deploy secure, scalable AI solutions that help companies move from basic productivity tools to transformative business intelligence.Topics Include:Introduction to GleanGlean targets Global 2000 companies for AI transformationEnterprise AI needs company context: data, people, processesBottom-up approach: deploy to all employees firstFocus on business results, not just productivity gainsGlean Assistant provides daily AI tool for employeesGlean Agents platform enables natural language agent buildingOpen APIs export context to enterprise systemsStarted as enterprise search, evolved to knowledge graphsKnowledge graphs map content, people, projects, and processesIndividual knowledge graphs created for each personGlean WorkAI platform includes search, protect, agentsGlean Protect ensures data security and AI governancePlatform integrates with existing enterprise tools nativelyMCP enables connection to various AI systemsStrong growth: $100M ARR, $700M+ funding raisedAWS partnership provides models, security, and deploymentParticipants:Matt “Kix” Kixmoeller – Chief Marketing Officer, GleanFurther Links:Website: https://www.glean.com/Glean on AWS MarketplaceSee how Amazon Web Services gives you the freedom to migrate, innovate, and scale your software company at https://aws.amazon.com/isv/

Legal Marketing Minutes with Nancy Myrland
073: What Is Ambient AI and Why Does It Matter To Lawyers?

Legal Marketing Minutes with Nancy Myrland

Play Episode Listen Later Jul 7, 2025 9:40 Transcription Available


What is ambient AI, and why does it matter to lawyers, law firms, and legal marketing and business development professionals? In this episode of Legal Marketing Minutes, I define ambient AI in clear terms and explain how it's quietly showing up in tools professionals already use. You'll hear a real-world example from another profession that illustrates the potential for this technology, and I'll walk you through what this shift means for your work in legal. We'll also discuss OpenAI's collaboration with renowned iPhone designer Jony Ive, what he and Sam Altman are building behind the scenes, and how it could impact the way we interact with AI in the future. Topics covered include what ambient AI is, where it's showing up in the legal profession, why it matters, and what ethical questions and compliance considerations law firms need to keep in mind. Ambient AI isn't just coming. It's already here. The question is whether your firm is ready. If you are in a place where you can leave a comment, please do so, as I would love to hear from you. If not, feel free to email me at nancy@myrlandmarketing.com. Also, my website, where you can find all of my contact information, and my other podcast, Legal Marketing Moments, can be found at https://myrlandmarketing.com/podcasts/legalmarketingminutes.com Thanks for spending a few of your Legal Marketing Minutes with me! If I can help you in your AI discernment and strategy, please let me know.