Occupation of loading and unloading ships
POPULARITY
Categories
Wes and Scott talk about their evolving home-server setups—Synology rigs, Mac minis, Docker vs. VMs, media servers, backups, Cloudflare Tunnels, and the real-world pros and cons of running your own hardware. Show Notes 00:00 Welcome to Syntax! 01:35 Why use a home server? 07:29 Apps for home servers 16:23 Home server hardware 18:27 Brought to you by Sentry.io 20:45 VMs vs containers and choosing the right software 25:53 How to expose services to the internet safely 30:38 Securing access to your server Hit us up on Socials! Syntax: X Instagram Tiktok LinkedIn Threads Wes: X Instagram Tiktok LinkedIn Threads Scott: X Instagram Tiktok LinkedIn Threads Randy: X Instagram YouTube Threads
Az előfizetők (de csak a Belső kör és Közösség csomagok tulajdonosai!) már szombat hajnalban hozzájutnak legfrissebb epizódunk teljes verziójához. A hétfőn publikált, ingyen meghallgatható verzió tíz perccel rövidebb. Itt írtunk arról, hogy tudod meghallgatni a teljes adást. Mohamedek között, Keith Richards és Yves Saint Laurent nyomában. Igazság és hitelesség a belső sávban. Nyelvtudomány és esztétikai terror. Pápázás. Tusfürdőmaffia a siralomházban. 00:23 Mohamedek között fagyláros havában. Black novemberre Alza-napok. Ez már a copywritereknek is sok.06:21 Keith Richards és a beat-generáció Marrakesben. Egykor érdekes helyek elturistásodása. Rövidnadrágban 20 fokban.12:47 A Docker robogó. Marrakes gyalog. Marrakes szezonon kívül.16:50 Bipoláris állapotok és stabilitás.20:29 A marokkói-amerikai kapcsolatok dicső múltja. Igazság és hitelesség Marokkóban.23:04 Olvasói levél: Siralomház. A Kisfogház Emlékhely.26:40 Olvasói levél: Kálmán László nyelvészeti elvei. Nádasdy Ádám és a szép napot. Fejes László és a hanti nyelv.32:17 Nem nyelvészet, stílus! Van, ami műveletlenség. Aki nem tud liberálisul, ne beszéljen liberálisul. Tamás Gáspár Miklós, mint stílusrendőr.40:30 Olvasói levél: jobbra tartás és kategorikus imperatívusz. Amikor Barangó keresztényeket irtott karácsonykor. A hat ember, aki élőben hallotta. Barangó: Nincs mentségem. M. Giorgio Richárd. XIV. Leó és a palesztin állam. A katolikus egyház baloldalisága Észak- és Dél-Amerikában.47:05 A nemzetközi tusfürdő-maffia.50:05 A marokkó.See omnystudio.com/listener for privacy information.
Andrew kicks off at 8 a.m. with six Docker containers running, and he and Chris dive into what modern Rails development looks like inside dev containers—covering Rails' own images and features, using Mise and Playwright instead of Selenium, and why OrbStack has replaced Docker Desktop on their Macs. They talk through the trade-offs of running services in containers, the quirks of Kamal's new local registry, and how Chris is turning all of this into a practical SaaS building series that shows real-world deployment and scaling with tools like JudoScale. Along the way, they weave in life updates about new babies, daycare costs, and even the power needs of AI data centers and nuclear energy. Press download now to hear more! LinksChris Oliver XAndrew Mason BlueskyJudoscale- Remote Ruby listener giftWhy Playwright Is Less Flaky Than Selenium by Justin SearlsRails Dev Container Images & FeaturesRuby on MacJudoscale-Process Utilization: How We Actually Track ThatGoRails- Domain Monitor SaaS- Adding the Domain ModelCheeky Pint PodcastSmarter Every Day (YouTube)The DiplomatThe Girlfriend Chris Oliver X/Twitter Andrew Mason X/Twitter Jason Charnes X/Twitter
Si parla degli acquisti di Federico durante il Black Friday e di un videocorso completo su Docker di Manuel Zavatta (che potete ottenere GRATIS).
In this episode of the Ardan Labs Podcast, Bill Kennedy talks with Caleb Mattingly, Founder and CEO of Secure Cloud Innovations, about his journey through cybersecurity, compliance, and entrepreneurship. Caleb shares insights into navigating complex compliance frameworks, the importance of vulnerability management, and building trust in the cybersecurity space. Beyond tech, he discusses his passion for linguistics, communication, and music—and how personal interests shape professional growth. The conversation also explores Caleb's entrepreneurial story, from a chance encounter at a swing dance club to building a thriving business during COVID-19, highlighting lessons in resilience, niche marketing, and the value of relationships in business.00:00 Introduction01:50 Cybersecurity and Compliance05:56 Vulnerability Management19:39 Education and Career Exploration26:35 Linguistics and Language Learning36:42 College Life and Personal Growth40:15 Music, Hobbies, and Self-Expression55:51 Balancing Work and Love01:12:08 Entering Cybersecurity01:23:05 Career Changes and New Beginnings01:26:49 Founding Secure Cloud Innovations01:39:56 Building Trust and Customer RelationshipsConnect with Caleb: LinkedIn: https://www.linkedin.com/in/caleb-h-mattingly/Mentioned in this Episode:Secure Cloud Innovations: https://trysci.co/Want more from Ardan Labs? You can learn Go, Kubernetes, Docker & more through our video training, live events, or through our blog!Online Courses : https://ardanlabs.com/education/ Live Events : https://www.ardanlabs.com/live-training-events/ Blog : https://www.ardanlabs.com/blog Github : https://github.com/ardanlabs
El servicio de almacenamiento en la nube del que te hablé hace unas semanas puede ser el destino de ciertas copias de seguridad de nuestro NAS, pero no directamente. Hay que currárselo un poco y levantar por el medio un servidor WebDAV con Docker. Te explico cómo.
In Folge 14 von TechTumult sprechen Stefan, Simon und Daniel (Apfelcast) darüber, wie du KI nicht nur über die Cloud nutzt, sondern direkt bei dir zu Hause – mit Mac mini, Docker, Home Assistant und Open-Source-LLMs. Ihr besprecht Vor- und Nachteile von Selfhosting, Datenschutz, Hardware-Auswahl und warum lokale KI in Zukunft für Creator und Unternehmen so spannend ist.
JupyterLite, a fully browser-based distribution of JupyterLab, is enabling new levels of global scalability in technical education. Developed by Sylvain Corlay's QuantStack team, it allows math and programming lessons to run entirely in students' browsers — kernel included — without relying on Docker or cloud-scale infrastructure. Its most prominent success is Capytale, a French national deployment that supports half a million high school students and over 200,000 weekly sessions from essentially a single server, which hosts only teaching content while computation happens locally in each browser.QuantStack, founded in 2016 as what Corlay calls an “accidental startup,” has since grown into a 30-person team contributing across Jupyter, Conda-Forge, and Apache Arrow. But JupyterLite embodies its most ambitious goal: making programming education accessible to countries with rapidly growing youth populations, such as Nigeria, where traditional cloud-hosted notebooks are impractical. Achieving a billion-user future will require advances in accessibility, collaboration, and expanding browser-based package support — efforts that depend on grants and foundation backing.Learn more from The New Stack about Project JupyterFrom Physics to the Future: Brian Granger on Project Jupyter in the Age of AIJupyter AI v3: Could It Generate an ‘Ecosystem of AI Personas?'Join our community of newsletter subscribers to stay on top of the news and at the top of your game. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
En este episodio profundizamos en la optimización de tus sistemas Docker. El objetivo de hoy: implementar nuevas y mejores prácticas para que tus contenedores sean más estables y eficientes. Si buscas cómo sacar el máximo partido a tu infraestructura autoalojada, esta es tu guía.Uno de los problemas más comunes al trabajar con Docker Compose es que los servicios dependientes (como una aplicación web o un servidor de correo) intentan conectarse a la base de datos antes de que esta haya terminado de arrancar, provocando errores.Te muestro mi solución PRO para esto: utilizamos la combinación de la directiva depends_on con la condición service_healthy.Esta configuración asegura que servicios críticos como Gitea (mi servicio de alojamiento de repositorios Git) y Stalwart (mi servidor de correo) solo se inician cuando su respectiva base de datos PostgreSQL ha pasado su chequeo de salud y está lista para aceptar conexiones. Esto garantiza una secuencia de inicio robusta y sin fallos, una mejora fundamental en la gestión de tus datos y sistemas.Gitea : Vemos cómo configurar el healthcheck para la base de datos PostgreSQL usando pg_isready y cómo el servicio Gitea espera por esta condición. También optimizamos el tráfico interno del runner de Gitea para que use la red interna de Docker (http://gitea:3000), reduciendo la carga de Traefik y mejorando la seguridad.Stalwart : En el caso de mi cliente de correo, he migrado la base de datos de RocketDB a PostgreSQL. La razón es sencilla: PostgreSQL es más transparente y me permite integrar sin esfuerzo mi contenedor personalizado (atareao/postgres-backup:latest) para hacer copias de seguridad eficientes y automatizadas.En este episodio, también te presento una nueva herramienta que me ha encantado: Dockpeek.Dockpeek es un panel de control autoalojado y muy ligero para Docker, perfecto para la gestión de contenedores en múltiples hosts. Si te gustan las herramientas que reemplazan funcionalidades complejas con soluciones sencillas, Dockpeek te va a encantar.Características destacadas: Acceso web con un clic, mapeo automático de puertos, registros de contenedores en vivo, integración con Traefik y chequeo de actualizaciones de imágenes.Te comparto el compose.yml que utilizo para instalar Dockpeek junto a Traefik.Quantum (Filebrowser): He ajustado los permisos y la configuración del servicio que utilizo para compartir archivos. Te explico la solución al problema de permisos que surgió al intentar usar un usuario que no es root, modificando el uid, gid y mode en la sección configs del compose.yml.Escucha el episodio para obtener el tutorial completo y adaptar estas soluciones a tu Raspberry Pi o VPS. ¡Es la forma más práctica de optimizar tu productividad y tus sistemas Linux!¡Suscríbete a "atareao con Linux" para no perderte ningún tutorial y llevar tu experiencia con Linux a un nivel PRO!
Join Wendy and Nate as they battle robot headaches, wrangle 3D printers, and bring tech holiday spirit to life! From migraine workarounds and sodium science, through epic 3D printing adventures (featuring OctoEverywhere!), to home automation, Docker disasters, and retro gaming resurrection, this episode is packed with open-source laughs and memorable tangents. Whether you love building robots or naming your Wi-Fi something wild, you'll find plenty of creative fuel—and team banter—in this jam-packed ride! Find the rest of the show notes at: https://tuxdigital.com/podcasts/linux-out-loud/lol-117/
Ya sea que gestiones una instancia de Syncthing, un backend de Obsidian con Docker, o tu proxy inverso con Traefik, sabes que la necesidad de revisar logs o reiniciar un contenedor puede surgir en cualquier momento. La solución habitual es la Terminal SSH, lo que te obliga a sacar el portátil o lidiar con interfaces incómodas en el móvil.En este episodio, te presento Docker Manager, una aplicación gratuita y open source construida con Flutter y un hermoso diseño Material Design. Esta herramienta es tu centro de comando definitivo para Docker, diseñado específicamente para pantallas pequeñas, permitiéndote abandonar el tedio del SSH para el 99% de las tareas diarias. Es una solución de productividad pura, muy en la línea de lo que buscamos en atareao con Linux: soluciones prácticas para "cualquier cosa que quieras hacer con Linux".Este episodio es un tutorial práctico paso a paso para que puedas poner Docker Manager en marcha y sacarle el máximo partido:Conexión Segura Multiserver: Explicaremos detalladamente cómo configurar la conexión a múltiples hosts Linux (VPS, Raspberry Pi, máquinas virtuales) y por qué debes utilizar la autenticación por clave privada SSH para mantener tu infraestructura segura. La app se integra perfectamente con tu pila de red móvil, lo que significa que funciona sin problemas a través de VPNs como WireGuard o Tailscale.Control Total de Contenedores: La facilidad para realizar operaciones esenciales: Start, Stop, Restart, Inspect y Remove con un solo toque. Haremos hincapié en el filtrado por Docker Compose Stacks, esencial para quien gestiona múltiples servicios como bases de datos o instancias de Rust alojadas en contenedores.Diagnóstico Avanzado en Movimiento:Logs en Vivo: Revisar los logs en tiempo real es vital para el debugging de emergencia.Estadísticas del Contenedor: Ver el uso de CPU y memoria al instante para identificar cuellos de botella.Shell Interactivo: La característica estrella. Te mostraremos cómo iniciar un shell (bash) dentro de un contenedor o en el host Linux mismo. Esto te da la libertad de usar herramientas como redis-cli o revisar configuraciones rápidas sin abrir un cliente SSH.Mantenimiento y Limpieza del Sistema: Analizaremos la función System Cleanup (Pruning) para deshacernos de esas imágenes y volúmenes "colgantes" que roban espacio.Gestión de Imágenes, Redes y Volúmenes: Un vistazo a cómo la aplicación simplifica la visualización y gestión de estos componentes clave de Docker. Incluso hablaremos de la flexibilidad para configurar el Docker CLI Path, lo que abre la puerta a la gestión de Podman también.Docker Manager es una herramienta indispensable que libera tu escritorio Linux y te da el poder de administración en tu bolsillo. Ya no tendrás que interrumpir tu flujo de trabajo en Neovim o cerrar tu sesión de escritorio GNOME para hacer una comprobación rápida. Es la solución perfecta para mantener tus servicios (desde un servidor web hasta una instancia de Obsidian) funcionando sin problemas 24/7.Más información y enlaces en las notas del episodio
In this episode of the Ardan Labs Podcast, Bill Kennedy talks with Catherine Johnson, VP of Global Solutions Engineering at Hydrolix, about her career journey through data management, scalability, and innovation. With a background spanning leadership roles at Hydrolix, Grafana Labs, and Oracle, Catherine shares insights into balancing engineering, entrepreneurship, and mentorship. She discusses the evolution of real-time data systems, the economics of storage, and the importance of data-driven decision-making. Catherine also opens up about taking a break from tech to teach dance, and how her passion for innovation and continuous learning fuels her leadership today.00:00 Introduction03:03 Data Management and Scalability05:58 Explosion of Data and Storage Needs09:01 Real-Time Data in Business14:49 Economics of Data Storage20:41 Education and Early Career31:09 Career Transitions and Growth46:10 Teaching Dance and Finding Balance53:16 Returning to Tech at Oracle01:08:10 Joining Elastic and Facing Burnout01:15:45 Leadership and Innovation at Hydrolix01:27:07 AI in Tech and Its LimitationsConnect with Catherine: LinkedIn: https://www.linkedin.com/in/catjopdx/Mentioned in this Episode:Hydrolix: https://hydrolix.io/Want more from Ardan Labs? You can learn Go, Kubernetes, Docker & more through our video training, live events, or through our blog!Online Courses : https://ardanlabs.com/education/ Live Events : https://www.ardanlabs.com/live-training-events/ Blog : https://www.ardanlabs.com/blog Github : https://github.com/ardanlabs
Te hablo de un contenedor Docker que permite manejar los archivos de nuestro servidor desde una simple, pero potente, interfaz web. El día que deje atrás mi Synology, perderé su maravilloso escritorio web y también la aplicación File Station, con lo que he buscado ya una alternativa a la altura.
Oscar Wieman bouwde ooit Minecraft-servers voor de lol, nu sleutelt hij aan de infrastructuur van de IND. Aan tafel met Randal en Annelies duikt hij in containers, Kubernetes en het automatiseren van deploys. Waarom werkt het “on my machine” wél en elders niet? Hoe voorkom je dat developers eindeloos zipjes mailen? En wat doe je als de overheid meer regeltjes heeft dan jij testomgevingen (of juist niet)? Shownotes It works on my machine… Fosdem.org Advertentie: Rabobank ITWil je weten hoe IT-professionals bij Rabobank stappen maken richting duurzamer IT-gebruik? Bekijk https://rabobank.jobs/IT Tijdschema0:00:00 Wie is Oscar en wat doet ‘ie bij de IND?0:04:30 DevOps, sysadmin of vliegende kiep?0:10:00 Containerisatie, Docker en dat soort dingen0:16:20 Waarom Kubernetes nodig werd0:22:40 Oscars route van hobby naar professie0:28:10 Overheidsprojecten en bureaucratische tape0:35:00 CI/CD, Git en containers bouwen0:42:00 “Works on my machine” en hoe Docker dat oplost0:50:00 Real-life use case: Rabobank-app en microservices0:58:00 Wat draait er allemaal achter de schermen bij de IND? #devops #docker #kubernetes #containers #infra #platformengineering #opensource #automation #overheid #MNOTSee omnystudio.com/listener for privacy information.
Dans cet épisode, Emmanuel, Katia et Guillaume discutent de Spring 7, Quarkus, d'Infinispan et Keycloak. On discute aussi de projets sympas comme Javelit, de comment démarre une JVM, du besoin d'argent de NTP. Et puis on discute du changement de carrière d'Emmanuel. Enregistré le 14 novembre 2025 Téléchargement de l'épisode LesCastCodeurs-Episode-332.mp3 ou en vidéo sur YouTube. News Emmanuel quitte Red Hat après 20 ans https://emmanuelbernard.com/blog/2025/11/13/leaving-redhat/ Langages Support HTTP/3 dans le HttpClient de JDK 26 - https://inside.java/2025/10/22/http3-support/ JDK 26 introduit le support de HTTP/3 dans l'API HttpClient existante depuis Java 11 HTTP/3 utilise le protocole QUIC sur UDP au lieu de TCP utilisé par HTTP/2 Par défaut HttpClient préfère HTTP/2, il faut explicitement configurer HTTP/3 avec Version.HTTP_3 Le client effectue automatiquement un downgrade vers HTTP/2 puis HTTP/1.1 si le serveur ne supporte pas HTTP/3 On peut forcer l'utilisation exclusive de HTTP/3 avec l'option H3_DISCOVERY en mode HTTP_3_URI_ONLY HttpClient apprend qu'un serveur supporte HTTP/3 via le header alt-svc (RFC 7838) et utilise cette info pour les requêtes suivantes La première requête peut utiliser HTTP/2 même avec HTTP/3 préféré, mais la seconde utilisera HTTP/3 si le serveur l'annonce L'équipe OpenJDK encourage les tests et retours d'expérience sur les builds early access de JDK 26 Librairies Eclispe Jetty et CometD changent leurs stratégie de support https://webtide.com/end-of-life-changes-to-eclipse-jetty-and-cometd/ À partir du 1er janvier 2026, Webtide ne publiera plus Jetty 9/10/11 et CometD 5/6/7 sur Maven Central Pendant 20 ans, Webtide a financé les projets Jetty et CometD via services et support, publiant gratuitement les mises à jour EOL Le comportement des entreprises a changé : beaucoup cherchent juste du gratuit plutôt que du véritable support Des sociétés utilisent des versions de plus de 10 ans sans migrer tant que les correctifs CVE sont gratuits Cette politique gratuite a involontairement encouragé la complaisance et retardé les migrations vers versions récentes MITRE développe des changements au système CVE pour mieux gérer les concepts d'EOL Webtide lance un programme de partenariat avec TuxCare et HeroDevs pour distribuer les résolutions CVE des versions EOL Les binaires EOL seront désormais distribués uniquement aux clients commerciaux et via le réseau de partenaires Webtide continue le support standard open-source : quand Jetty 13 sortira, Jetty 12.1 recevra des mises à jour pendant 6 mois à un an Ce changement vise à clarifier la politique EOL avec une terminologie industrielle établie Améliorations cloud du SDK A2A Java https://quarkus.io/blog/quarkus-a2a-cloud-enhancements/ Version 0.3.0.Final du SDK A2A Java apporte des améliorations pour les environnements cloud et distribués Composants en mémoire remplacés par des implémentations persistantes et répliquées pour environnements multi-instances JpaDatabaseTaskStore et JpaDatabasePushNotificationConfigStore permettent la persistance des tâches et configurations en base PostgreSQL ReplicatedQueueManager assure la réplication des événements entre instances A2A Agent via Kafka et MicroProfile Reactive Messaging Exemple complet de déploiement Kubernetes avec Kind incluant PostgreSQL, Kafka via Strimzi, et load balancing entre pods Démonstration pratique montrant que les messages peuvent être traités par différents pods tout en maintenant la cohérence des tâches Architecture inspirée du SDK Python A2A, permettant la gestion de tâches asynchrones longues durée en environnement distribué Quarkus 3.29 sort avec des backends de cache multiples et support du débogueur Qute https://quarkus.io/blog/quarkus-3-29-released/ Possibilité d'utiliser plusieurs backends de cache simultanément dans une même application Chaque cache peut être associé à un backend spécifique (par exemple Caffeine et Redis ou Infinispan) Support du Debug Adapter Protocol (DAP) pour déboguer les templates Qute directement dans l'IDE et dans la version 3.28 Configuration programmatique de la protection CSRF via une API fluent Possibilité de restreindre les filtres OIDC à des flux d'authentification spécifiques avec annotations Support des dashboards Grafana personnalisés via fichiers JSON dans META-INF/grafana/ Extension Liquibase MongoDB supporte désormais plusieurs clients simultanés Amélioration significative des performances de build avec réduction des allocations mémoire Parallélisation de tâches comme la génération de proxies Hibernate ORM et la construction des Jar Et l'utilisation des fichiers .proto est plus simple dans Quarkus avbec Quarkus gRPC Zero https://quarkus.io/blog/grpc-zero/ c'est toujours galere des fichiers .proto car les generateurs demandent des executables natifs maintenant ils sont bundlés dans la JVM et vous n'avez rien a configurer cela utilise Caffeine pour faire tourner cela en WASM dans la JVM Spring AI 1.1 est presque là https://spring.io/blog/2025/11/08/spring-ai-1-1-0-RC1-available-now support des MCP tool caching pour les callback qui reduit les iooerations redondantes Access au contenu de raisonnement OpenAI Un modele de Chat MongoDB Support du modele de penser Ollama Reessaye sur les echec de reseau OpenAI speech to text Spring gRPC Les prochaines étapes pour la 1.0.0 https://spring.io/blog/2025/11/05/spring-grpc-next-steps Spring gRPC 1.0 arrive prochainement avec support de Spring Boot 4 L'intégration dans Spring Boot 4.0 est reportée, prévue pour Spring Boot 4.1 Les coordonnées Maven restent sous org.springframework.grpc pour la version 1.0 Le jar spring-grpc-test est renommé en spring-grpc-test-spring-boot-autoconfigure Les packages d'autoconfiguration changent de nom nécessitant de modifier les imports Les dépendances d'autoconfiguration seront immédiatement dépréciées après la release 1.0 Migration minimale attendue pour les projets utilisant déjà la version 0.x La version 1.0.0-RC1 sera publiée dès que possible avant la version finale Spring arrete le support reactif d'Apache Pulsar https://spring.io/blog/2025/10/29/spring-pulsar-reactive-discontinued logique d'évaluer le temps passé vs le nombre d'utilisateurs c'est cependant une tendance qu'on a vu s'accélerer Spring 7 est sorti https://spring.io/blog/2025/11/13/spring-framework-7-0-general-availability Infrastructure Infinispan 16.0 https://infinispan.org/blog/2025/11/10/infinispan-16-0 Ajout majeur : migration en ligne sans interruption pour les nœuds d'un cluster (rolling upgrades) (infinispan.org) Messages de clustering refaits avec Protocol Buffers + ProtoStream : meilleure compatibilité, schéma évolutif garanti (infinispan.org) Console Web améliorée API dédiée de gestion des schémas (SchemasAdmin) pour gérer les schémas ProtoStream à distance (infinispan.org) Module de requête (query) optimisé : support complet des agrégations (sum, avg …) dans les requêtes indexées en cluster grâce à l'intégration de Hibernate Search 8.1 (infinispan.org) Serveur : image conteneur minimalisée pour réduire la surface d'attaque (infinispan.org) démarrage plus rapide grâce à séparation du démarrage cache/serveur (infinispan.org) caches pour connecteurs (Memcached, RESP) créés à la demande (on-demand) et non à l'initiaton automatique (infinispan.org) moteur Lua 5.1 mis à jour avec corrections de vulnérabilités et opérations dangereuses désactivées (infinispan.org) Support JDK : version minimale toujours JDK 17 (infinispan.org) prise en charge des threads virtuels (virtual threads) et des fonctionnalités AOT (Ahead-of-Time) de JDK plus récentes (infinispan.org) Web Javelit, une nouvelle librairie Java inspirée de Streamlit pour faire facilement et rapidement des petites interfaces web https://glaforge.dev/posts/2025/10/24/javelit-to-create-quick-interactive-app-frontends-in-java/ Site web du projet : https://javelit.io/ Javelit : outil pour créer rapidement des applications de données (mais pas que) en Java. Simplifie le développement : élimine les tracas du frontend et de la gestion des événements. Transforme une classe Java en application web en quelques minutes. Inspiré par la simplicité de Streamlit de l'écosystème Python (ou Gradio et Mesop), mais pour Java. Développement axé sur la logique : pas de code standard répétitif (boilerplate), rechargement à chaud. Interactions faciles : les widgets retournent directement leur valeur, sans besoin de HTML/CSS/JS ou gestion d'événements. Déploiement flexible : applications autonomes ou intégrables dans des frameworks Java (Spring, Quarkus, etc.). L'article de Guillaume montre comment créer une petite interface pour créer et modifier des images avec le modèle génératif Nano Banana Un deuxième article montre comment utiliser Javelit pour créer une interface de chat avec LangChain4j https://glaforge.dev/posts/2025/10/25/creating-a-javelit-chat-interface-for-langchain4j/ Améliorer l'accessibilité avec les applis JetPack Compose https://blog.ippon.fr/2025/10/29/rendre-son-application-accessible-avec-jetpack-compose/ TalkBack est le lecteur d'écran Android qui vocalise les éléments sélectionnés pour les personnes malvoyantes Accessibility Scanner et les outils Android Studio détectent automatiquement les problèmes d'accessibilité statiques Les images fonctionnelles doivent avoir un contentDescription, les images décoratives contentDescription null Le contraste minimum requis est de 4.5:1 pour le texte normal et 3:1 pour le texte large ou les icônes Les zones cliquables doivent mesurer au minimum 48dp x 48dp pour faciliter l'interaction Les formulaires nécessitent des labels visibles permanents et non de simples placeholders qui disparaissent Modifier.semantics permet de définir l'arbre sémantique lu par les lecteurs d'écran Les propriétés mergeDescendants et traversalIndex contrôlent l'ordre et le regroupement de la lecture Diriger le navigateur Chrome avec le modèle Gemini Computer Use https://glaforge.dev/posts/2025/11/03/driving-a-web-browser-with-gemini-computer-use-model-in-java/ Objectif : Automatiser la navigation web en Java avec le modèle "Computer Use" de Gemini 2.5 Pro. Modèle "Computer Use" : Gemini analyse des captures d'écran et génère des actions d'interface (clic, saisie, etc.). Outils : Gemini API, Java, Playwright (pour l'interaction navigateur). Fonctionnement : Boucle agent où Gemini reçoit une capture, propose une action, Playwright l'exécute, puis une nouvelle capture est envoyée à Gemini. Implémentation clé : Toujours envoyer une capture d'écran à Gemini après chaque action pour qu'il comprenne l'état actuel. Défis : Lenteur, gestion des CAPTCHA et pop-ups (gérables). Potentiel : Automatisation des tâches web répétitives, création d'agents autonomes. Data et Intelligence Artificielle Apicurio ajoute le support de nouveaux schema sans reconstruire Apicurio https://www.apicur.io/blog/2025/10/27/custom-artifact-types Apicurio Registry 3.1.0 permet d'ajouter des types d'artefacts personnalisés au moment du déploiement sans recompiler le projet Supporte nativement OpenAPI, AsyncAPI, Avro, JSON Schema, Protobuf, GraphQL, WSDL et XSD Trois approches d'implémentation disponibles : classes Java pour la performance maximale, JavaScript/TypeScript pour la facilité de développement, ou webhooks pour une flexibilité totale Configuration via un simple fichier JSON pointant vers les implémentations des composants personnalisés Les scripts JavaScript sont exécutés via QuickJS dans un environnement sandboxé sécurisé Un package npm TypeScript fournit l'autocomplétion et la sécurité de type pour le développement Six composants optionnels configurables : détection automatique de type, validation, vérification de compatibilité, canonicalisation, déréférencement et recherche de références Cas d'usage typiques : formats propriétaires internes, support RAML, formats legacy comme WADL, schémas spécifiques à un domaine métier Déploiement simple via Docker en montant les fichiers de configuration et scripts comme volumes Les performances varient selon l'approche : Java offre les meilleures performances, JavaScript un bon équilibre, webhooks la flexibilité maximale Le truc interessant c'est que c'est Quarkus based et donc demandait le rebuilt donc pour eviter cela, ils ont ajouter QuickJS via Chicorey un moteur WebAssembly GPT 5.1 pour les développeurs est sorti. https://openai.com/index/gpt-5-1-for-developers/ C'est le meilleur puisque c'est le dernier :slightly_smiling_face: Raisonnement Adaptatif et Efficace : GPT-5.1 ajuste dynamiquement son temps de réflexion en fonction de la complexité de la tâche, le rendant nettement plus rapide et plus économique en jetons pour les tâches simples, tout en maintenant des performances de pointe sur les tâches difficiles. Nouveau Mode « Sans Raisonnement » : Un mode (reasoning_effort='none') a été introduit pour les cas d'utilisation sensibles à la latence, permettant une réponse plus rapide avec une intelligence élevée et une meilleure exécution des outils. Cache de Prompt Étendu : La mise en cache des invites est étendue jusqu'à 24 heures (contre quelques minutes auparavant), ce qui réduit la latence et le coût pour les interactions de longue durée (chats multi-tours, sessions de codage). Les jetons mis en cache sont 90 % moins chers. Améliorations en Codage : Le modèle offre une meilleure personnalité de codage, une qualité de code améliorée et de meilleures performances sur les tâches d'agenticité de code, atteignant 76,3 % sur SWE-bench Verified. Nouveaux Outils pour les Développeurs : Deux nouveaux outils sont introduits ( https://cookbook.openai.com/examples/build_a_coding_agent_with_gpt-5.1 ) : L'outil apply_patch pour des modifications de code plus fiables via des diffs structurés. L'outil shell qui permet au modèle de proposer et d'exécuter des commandes shell sur une machine locale, facilitant les boucles d'inspection et d'exécution. Disponibilité : GPT-5.1 (ainsi que les modèles gpt-5.1-codex) est disponible pour les développeurs sur toutes les plateformes API payantes, avec les mêmes tarifs et limites de débit que GPT-5. Comparaison de similarité d'articles et de documents avec les embedding models https://glaforge.dev/posts/2025/11/12/finding-related-articles-with-vector-embedding-models/ Principe : Convertir les articles en vecteurs numériques ; la similarité sémantique est mesurée par la proximité de ces vecteurs. Démarche : Résumé des articles via Gemini-2.5-flash. Conversion des résumés en vecteurs (embeddings) par Gemini-embedding-001. Calcul de la similarité entre vecteurs par similarité cosinus. Affichage des 3 articles les plus pertinents (>0.75) dans le frontmatter Hugo. Bilan : Approche "résumé et embedding" efficace, pragmatique et améliorant l'engagement des lecteurs. Outillage Composer : Nouveau modèle d'agent rapide pour l'ingénierie logicielle - https://cursor.com/blog/composer Composer est un modèle d'agent conçu pour l'ingénierie logicielle qui génère du code quatre fois plus rapidement que les modèles similaires Le modèle est entraîné sur de vrais défis d'ingénierie logicielle dans de grandes bases de code avec accès à des outils de recherche et d'édition Il s'agit d'un modèle de type mixture-of-experts optimisé pour des réponses interactives et rapides afin de maintenir le flux de développement L'entraînement utilise l'apprentissage par renforcement dans divers environnements de développement avec des outils comme la lecture de fichiers, l'édition, les commandes terminal et la recherche sémantique Cursor Bench est un benchmark d'évaluation basé sur de vraies demandes d'ingénieurs qui mesure la correction et le respect des abstractions du code existant Le modèle apprend automatiquement des comportements utiles comme effectuer des recherches complexes, corriger les erreurs de linter et écrire des tests unitaires L'infrastructure d'entraînement utilise PyTorch et Ray avec des kernels MXFP8 pour entraîner sur des milliers de GPUs NVIDIA Le système exécute des centaines de milliers d'environnements de codage sandboxés concurrents dans le cloud pour l'entraînement Composer est déjà utilisé quotidiennement par les développeurs de Cursor pour leur propre travail Le modèle se positionne juste derrière GPT-5 et Sonnet 4.5 en termes de performance sur les benchmarks internes Rex sur l'utilisation de l'IA pour les développeurs, un gain de productivité réel et des contextes adaptés https://mcorbin.fr/posts/2025-10-17-genai-dev/ Un développeur avec 18 ans d'expérience partage son retour sur l'IA générative après avoir changé d'avis Utilise exclusivement Claude Code dans le terminal pour coder en langage naturel Le "vibe coding" permet de générer des scripts et interfaces sans regarder le code généré Génération rapide de scripts Python pour traiter des CSV, JSON ou créer des interfaces HTML Le mode chirurgien résout des bugs complexes en one-shot, exemple avec un plugin Grafana fixé en une minute Pour le code de production, l'IA génère les couches repository, service et API de manière itérative, mais le dev controle le modele de données Le développeur relit toujours le code et ajuste manuellement ou via l'IA selon le besoin L'IA ne remplacera pas les développeurs car la réflexion, conception et expertise technique restent essentielles La construction de produits robustes, scalables et maintenables nécessite une expérience humaine L'IA libère du temps sur les tâches répétitives et permet de se concentrer sur les aspects complexes ce que je trouve interessant c'est la partie sur le code de prod effectivement, je corrige aussi beaucoup les propositions de l'IA en lui demandant de faire mieux dans tel ou tel domaine Sans guide, tout cela serait perdu Affaire a suivre un article en parallele sur le métier de designer https://blog.ippon.fr/2025/11/03/lia-ne-remplace-pas-un-designer-elle-amplifie-la-difference-entre-faire-et-bien-faire/ Plus besoin de se rappeler les racourcis dans IntelliJ idea avec l'universal entry point https://blog.jetbrains.com/idea/2025/11/universal-entry-point-a-single-entry-point-for-context-aware-coding-assistance/ IntelliJ IDEA introduit Command Completion, une nouvelle façon d'accéder aux actions de l'IDE directement depuis l'éditeur Fonctionne comme la complétion de code : tapez point (.) pour voir les actions contextuelles disponibles Tapez double point (..) pour filtrer et n'afficher que les actions disponibles Propose des corrections, refactorings, génération de code et navigation selon le contexte Complète les fonctionnalités existantes sans les remplacer : raccourcis, Alt+Enter, Search Everywhere Facilite la découverte des fonctionnalités de l'IDE sans interrompre le flux de développement En Beta dans la version 2025.2, sera activé par défaut dans 2025.3 Support actuel pour Java et Kotlin, avec actions spécifiques aux frameworks comme Spring et Hibernate Homebrew, package manage pour macOS et Linux passe en version 5 https://brew.sh/2025/11/12/homebrew-5.0.0/ Téléchargements Parallèles par Défaut : Le paramètre HOMEBREW_DOWNLOAD_CONCURRENCY=auto est activé par défaut, permettant des téléchargements concurrents pour tous les utilisateurs, avec un rapport de progression. Support Linux ARM64/AArch64 en Tier 1 : Le support pour Linux ARM64/AArch64 a été promu au niveau "Tier 1" (support officiel de premier plan). Feuille de Route pour les Dépréciations macOS : Septembre 2026 (ou plus tard) : Homebrew ne fonctionnera plus sur macOS Catalina (10.15) et versions antérieures. macOS Intel (x86_64) passera en "Tier 3" (fin du support CI et des binaires précompilés/bottles). Septembre 2027 (ou plus tard) : Homebrew ne fonctionnera plus sur macOS Big Sur (11) sur Apple Silicon ni du tout sur Intel (x86_64). Sécurité et Casks : Dépréciation des Casks sans signature de code. Désactivation des Casks échouant aux vérifications Gatekeeper en septembre 2026. Les options --no-quarantine et --quarantine sont dépréciés pour ne plus faciliter le contournement des fonctionnalités de sécurité de macOS. Nouvelles Fonctionnalités & Améliorations : Support officiel pour macOS 26 (Tahoe). brew bundle supporte désormais l'installation de packages Go via un Brewfile. Ajout de la commande brew info --sizes pour afficher la taille des formulae et casks. La commande brew search --alpine permet de chercher des packages Alpine Linux. Architecture Selon l'analyste RedMonk, Java reste très pertinent dans l'aire de l'IA et des agents https://redmonk.com/jgovernor/java-relevance-in-the-ai-era-agent-frameworks-emerge/ Java reste pertinent à l'ère de l'IA, pas besoin d'apprendre une pile technique entièrement nouvelle. Capacité d'adaptation de Java ("anticorps") aux innovations (Big Data, cloud, IA), le rendant idéal pour les contextes d'entreprise. L'écosystème JVM offre des avantages sur Python pour la logique métier et les applications sophistiquées, notamment en termes de sécurité et d'évolutivité. Embabel (par Rod Johnson, créateur de Spring) : un framework d'agents fortement typé pour JVM, visant le déterminisme des projets avant la génération de code par LLM. LangChain4J : facilite l'accès aux capacités d'IA pour les développeurs Java, s'aligne sur les modèles d'entreprise établis et permet aux LLM d'appeler des méthodes Java. Koog (Jetbrains) : framework d'agents basé sur Kotlin, typé et spécifique aux développeurs JVM/Kotlin. Akka : a pivoté pour se concentrer sur les flux de travail d'agents IA, abordant la complexité, la confiance et les coûts des agents dans les systèmes distribués. Le Model Context Protocol (MCP) est jugé insuffisant, manquant d'explicabilité, de découvrabilité, de capacité à mélanger les modèles, de garde-fous, de gestion de flux, de composabilité et d'intégration sécurisée. Les développeurs Java sont bien placés pour construire des applications compatibles IA et intégrer des agents. Des acteurs majeurs comme IBM, Red Hat et Oracle continuent d'investir massivement dans Java et son intégration avec l'IA. Sécurité AI Deepfake, Hiring … A danger réel https://www.eu-startups.com/2025/10/european-startups-get-serious-about-deepfakes-as-ai-fraud-losses-surpass-e1-3-billion/ Pertes liées aux deepfakes en Europe : > 1,3 milliard € (860 M € rien qu'en 2025). Création de deepfakes désormais possible pour quelques euros. Fraudes : faux entretiens vidéo, usurpations d'identité, arnaques diverses. Startups actives : Acoru, IdentifAI, Trustfull, Innerworks, Keyless (détection et prévention). Réglementation : AI Act et Digital Services Act imposent transparence et contrôle. Recommandations : vérifier identités, former employés, adopter authentification multi-facteurs. En lien : https://www.techmonitor.ai/technology/cybersecurity/remote-hiring-cybersecurity 1 Candidat sur 4 sera Fake en 2028 selon Gartner research https://www.gartner.com/en/newsroom/press-releases/2025-07-31-gartner-survey-shows-j[…]-percent-of-job-applicants-trust-ai-will-fairly-evaluate-them Loi, société et organisation Amazon - prévoit supprimer 30.000 postes https://www.20minutes.fr/economie/4181936-20251028-amazon-prevoit-supprimer-30-000-emplois-bureau-selon-plusieurs-medias Postes supprimés : 30 000 bureaux Part des effectifs : ~10 % des employés corporatifs Tranche confirmée : 14 000 postes Divisions touchées : RH, Opérations, Devices & Services, Cloud Motifs : sur-recrutement, bureaucratie, automatisation/IA Accompagnement : 90 jours pour poste interne + aides Non concernés : entrepôts/logistique Objectif : concentrer sur priorités stratégiques NTP a besoin d'argent https://www.ntp.org/ Il n'est que le protocole qui synchronise toutes les machines du monde La fondation https://www.nwtime.org/ recherche 11000$ pour maintenir son activité Rubrique débutant Une plongée approfondie dans le démarrage de la JVM https://inside.java/2025/01/28/jvm-start-up La JVM effectue une initialisation complexe avant d'exécuter le code : validation des arguments, détection des ressources système et sélection du garbage collector approprié Le chargement de classes suit une stratégie lazy où chaque classe charge d'abord ses dépendances dans l'ordre de déclaration, créant une chaîne d'environ 450 classes même pour un simple Hello World La liaison de classes comprend trois sous-processus : vérification de la structure, préparation avec initialisation des champs statiques à leurs valeurs par défaut, et résolution des références symboliques du Constant Pool Le CDS améliore les performances au démarrage en fournissant des classes pré-vérifiées, réduisant le travail de la JVM L'initialisation de classe exécute les initialiseurs statiques via la méthode spéciale clinit générée automatiquement par javac Le Project Leyden introduit la compilation AOT dans JDK 24 pour réduire le temps de démarrage en effectuant le chargement et la liaison de classes en avance de phase Pas si débutant finalement Conférences La liste des conférences provenant de Developers Conferences Agenda/List par Aurélie Vache et contributeurs : 12-14 novembre 2025 : Devoxx Morocco - Marrakech (Morocco) 15-16 novembre 2025 : Capitole du Libre - Toulouse (France) 19 novembre 2025 : SREday Paris 2025 Q4 - Paris (France) 19-21 novembre 2025 : Agile Grenoble - Grenoble (France) 20 novembre 2025 : OVHcloud Summit - Paris (France) 21 novembre 2025 : DevFest Paris 2025 - Paris (France) 24 novembre 2025 : Forward Data & AI Conference - Paris (France) 27 novembre 2025 : DevFest Strasbourg 2025 - Strasbourg (France) 28 novembre 2025 : DevFest Lyon - Lyon (France) 1-2 décembre 2025 : Tech Rocks Summit 2025 - Paris (France) 4-5 décembre 2025 : Agile Tour Rennes - Rennes (France) 5 décembre 2025 : DevFest Dijon 2025 - Dijon (France) 9-11 décembre 2025 : APIdays Paris - Paris (France) 9-11 décembre 2025 : Green IO Paris - Paris (France) 10-11 décembre 2025 : Devops REX - Paris (France) 10-11 décembre 2025 : Open Source Experience - Paris (France) 11 décembre 2025 : Normandie.ai 2025 - Rouen (France) 14-17 janvier 2026 : SnowCamp 2026 - Grenoble (France) 22 janvier 2026 : DevCon #26 : sécurité / post-quantique / hacking - Paris (France) 29-31 janvier 2026 : Epitech Summit 2026 - Paris - Paris (France) 2-5 février 2026 : Epitech Summit 2026 - Moulins - Moulins (France) 2-6 février 2026 : Web Days Convention - Aix-en-Provence (France) 3 février 2026 : Cloud Native Days France 2026 - Paris (France) 3-4 février 2026 : Epitech Summit 2026 - Lille - Lille (France) 3-4 février 2026 : Epitech Summit 2026 - Mulhouse - Mulhouse (France) 3-4 février 2026 : Epitech Summit 2026 - Nancy - Nancy (France) 3-4 février 2026 : Epitech Summit 2026 - Nantes - Nantes (France) 3-4 février 2026 : Epitech Summit 2026 - Marseille - Marseille (France) 3-4 février 2026 : Epitech Summit 2026 - Rennes - Rennes (France) 3-4 février 2026 : Epitech Summit 2026 - Montpellier - Montpellier (France) 3-4 février 2026 : Epitech Summit 2026 - Strasbourg - Strasbourg (France) 3-4 février 2026 : Epitech Summit 2026 - Toulouse - Toulouse (France) 4-5 février 2026 : Epitech Summit 2026 - Bordeaux - Bordeaux (France) 4-5 février 2026 : Epitech Summit 2026 - Lyon - Lyon (France) 4-6 février 2026 : Epitech Summit 2026 - Nice - Nice (France) 12-13 février 2026 : Touraine Tech #26 - Tours (France) 26-27 mars 2026 : SymfonyLive Paris 2026 - Paris (France) 27-29 mars 2026 : Shift - Nantes (France) 31 mars 2026 : ParisTestConf - Paris (France) 16-17 avril 2026 : MiXiT 2026 - Lyon (France) 22-24 avril 2026 : Devoxx France 2026 - Paris (France) 23-25 avril 2026 : Devoxx Greece - Athens (Greece) 6-7 mai 2026 : Devoxx UK 2026 - London (UK) 22 mai 2026 : AFUP Day 2026 Lille - Lille (France) 22 mai 2026 : AFUP Day 2026 Paris - Paris (France) 22 mai 2026 : AFUP Day 2026 Bordeaux - Bordeaux (France) 22 mai 2026 : AFUP Day 2026 Lyon - Lyon (France) 17 juin 2026 : Devoxx Poland - Krakow (Poland) 11-12 juillet 2026 : DevLille 2026 - Lille (France) 4 septembre 2026 : JUG Summer Camp 2026 - La Rochelle (France) 17-18 septembre 2026 : API Platform Conference 2026 - Lille (France) 5-9 octobre 2026 : Devoxx Belgium - Antwerp (Belgium) Nous contacter Pour réagir à cet épisode, venez discuter sur le groupe Google https://groups.google.com/group/lescastcodeurs Contactez-nous via X/twitter https://twitter.com/lescastcodeurs ou Bluesky https://bsky.app/profile/lescastcodeurs.com Faire un crowdcast ou une crowdquestion Soutenez Les Cast Codeurs sur Patreon https://www.patreon.com/LesCastCodeurs Tous les épisodes et toutes les infos sur https://lescastcodeurs.com/
Andrew Lock is a staff software engineer at Datadog and educator whose contributions to the .NET ecosystem have shaped how developers approach modern web applications. Located in the UK, Andrew is a Microsoft MVP, Author of ASP.NET Core in Action, and has an active blog all about his experience working with .NET and ASP.NET Core. Topics of Discussion: [2:56] Andrew talks about appreciating the joy of coding and the minutiae of figuring out the correct way to do things. [3:28] Andrew discusses the various testing frameworks available for .NET, including MS Test, NUnit, XUnit, and TUnit. He explains the history and evolution of these frameworks, noting that XUnit has become the de facto default version. [7:41] Andrew explains his interest in TUnit, a newer testing library that addresses some of the limitations of XUnit. [9:29] TUnit is designed to be fast, supporting parallel execution and native AOT for better performance. [12:16] Is there a way to radically speed up the execution of big test suites? [15:39] Andrew explains the importance of each type of test in providing confidence that the software works as intended. [21:26] Andrew notes that full system tests can provide strong confidence by exercising critical pathways in the application. [29:44] Andrew mentions that tools like Octopus Deploy can be used to automate smoke tests as part of the deployment process. [30:26] Advice to new developers regarding automated testing, and the importance of writing code that is easy to test, and thinking about testing when writing code. Mentioned in this Episode: Clear Measure Way Architect Forum Software Engineer Forum Andrew Lock "Andrew Lock: Containers in .NET8 - Ep 281" "Andrew Lock: Web Applications in .NET6 - Ep 198" "Updates to Docker images in .NET8" Want to Learn More? Visit AzureDevOps.Show for show notes and additional episodes.
This show has been flagged as Explicit by the host. | Title | Author | Narrated By | Duration | Released | |----------------------|----------------------|----------------------|----------|------------| | Excession - Culture | Iain M. Banks | Peter Kenny | 15:55:00 | 2013-03-07 | | The Martian | Andy Weir | R. C. Bray | 10:53:00 | 2013-03-22 | | Alien: Out of the Sh | Tim Lebbon, Dirk Mag | Rutger Hauer, Corey | 04:28:00 | 2016-04-26 | | The Best Science Fic | Neil Clarke - editor | Amy Tallmadge, Jerem | 28:04:00 | 2016-06-07 | | Aurora: CV-01 - Fron | Ryk Brown | Jeffrey Kafer | 07:15:00 | 2013-05-13 | | The Rings of Haven - | Ryk Brown | Jeffrey Kafer | 05:45:00 | 2013-06-03 | | The Legend of Corina | Ryk Brown | Jeffrey Kafer | 06:39:00 | 2013-06-17 | | Freedom's Dawn - Fro | Ryk Brown | Jeffrey Kafer | 09:06:00 | 2013-07-08 | | Rise of the Corinari | Ryk Brown | Jeffrey Kafer | 10:54:00 | 2013-07-29 | | Head of the Dragon - | Ryk Brown | Jeffrey Kafer | 12:41:00 | 2013-08-19 | | The Expanse - The Fr | Ryk Brown | Jeffrey Kafer | 10:19:00 | 2013-12-10 | | Celestia CV-02 - The | Ryk Brown | Jeffrey Kafer | 09:00:00 | 2013-12-23 | | Resistance - Frontie | Ryk Brown | Jeffrey Kafer | 09:07:00 | 2014-01-20 | | Liberation - The Fro | Ryk Brown | Jeffrey Kafer | 10:40:00 | 2014-04-15 | | Monkey | Wu Ch'êng-ên, Arthur | Kenneth Williams | 13:39:00 | 2015-10-02 | | Artemis | Andy Weir | Rosario Dawson | 08:57:00 | 2017-11-14 | | Born of the Ashes - | Ryk Brown | Jeffrey Kafer | 12:00:00 | 2014-07-16 | | Rise of the Alliance | Ryk Brown | Jeffrey Kafer | 10:46:00 | 2015-01-20 | | A Show of Force - Fr | Ryk Brown | Jeffrey Kafer | 13:09:00 | 2015-06-23 | | Frontiers Saga Serie | Ryk Brown | Jeffrey Kafer | 10:20:00 | 2015-11-24 | | That Which Other Men | Ryk Brown | Jeffrey Kafer | 14:28:00 | 2016-03-29 | | Colorless Tsukuru Ta | Haruki Murakami | Michael Fenton Steve | 09:07:00 | 2014-08-12 | | Celtic Mythology: Cl | Scott Lewis | Oliver Hunt | 03:23:00 | 2018-07-18 | | Children of Dune | Frank Herbert | Scott Brick, Simon V | 16:51:00 | 2008-02-05 | | Dune | Frank Herbert | Scott Brick, Orlagh | 21:02:00 | 2006-12-31 | | Dune Messiah | Frank Herbert | Scott Brick, Katheri | 08:57:00 | 2007-10-01 | | Bandersnatch - C.S. | Diana Pavlac Glyer | Michael Ward | 06:29:00 | 2016-09-26 | | The Fighters | C. J. Chivers | Scott Brick | 13:45:00 | 2018-08-14 | | Masters of Doom - Ho | David Kushner | Wil Wheaton | 12:43:00 | 2012-07-12 | | Salvation - The Salv | Peter F. Hamilton | John Lee | 19:02:00 | 2018-09-06 | | Cibola Burn - Book 4 | James S. A. Corey | Jefferson Mays | 20:07:00 | 2015-05-07 | | Lost at Sea: The Jon | Jon Ronson | Jon Ronson | 15:22:00 | 2012-10-11 | | Data Science: The Ul | Herbert Jones | Sam Slydell | 05:18:00 | 2018-11-28 | | The Coen Brothers | Adam Nayman | Rob Shapiro | 09:55:00 | 2018-09-11 | | Nemesis Games - The | James S. A. Corey | Jefferson Mays | 18:06:00 | 2015-06-02 | | The Ten Types of Hum | Dexter Dias | Tom Clegg | 26:32:00 | 2017-07-06 | | Delta-v | Daniel Suarez | Jeff Gurner | 16:42:00 | 2019-04-23 | | God Emperor of Dune | Frank Herbert | Simon Vance | 15:48:00 | 2007-12-30 | | Dreaming in Code - T | Scott Rosenberg | Kyle McCarley | 12:01:00 | 2012-12-18 | | Ghost in the Wires - | Kevin Mitnick, Willi | Ray Porter | 13:59:00 | 2011-08-15 | | Gibraltar Sun - Gibr | Michael McCollum | Ramon De Ocampo | 10:05:00 | 2013-02-28 | | The Tragedy of King | William Shakespeare | full cast | 01:46:00 | 2009-08-28 | | Blind Faith | Ben Elton | Michael Maloney | 04:22:00 | 2007-11-07 | | Talking to Strangers | Malcolm Gladwell | Malcolm Gladwell | 08:42:00 | 2019-09-10 | | The Hidden Life of T | Peter Wohlleben | Mike Grady | 07:33:00 | 2016-09-13 | | Orcs | Stan Nicholls | John Lee | 24:43:00 | 2011-09-08 | | Behave | Robert M. Sapolsky | Michael Goldstrom | 26:27:00 | 2018-08-16 | | The City and the Sta | Arthur C. Clarke | Mike Grady | 09:42:00 | 2013-01-20 | | The Forbidden City - | Charles River Editor | Colin Fluxman | 01:13:00 | 2017-02-27 | | Foundation - The Fou | Isaac Asimov | William Hope | 08:56:00 | 2019-09-26 | | Children of the Mind | Orson Scott Card | Gabrielle de Cuir, J | 13:30:00 | 2004-08-04 | | Shahnameh - The Epic | Ferdowsi | Marc Thompson, Franc | 12:01:00 | 2017-12-22 | | The Cuckoo's Egg - T | Cliff Stoll | Will Damron | 12:46:00 | 2020-01-31 | | We the Living | Ayn Rand | Mary Woods | 18:01:00 | 2007-12-24 | | The Clock Mirage - O | Joseph Mazur | Keith Sellon-Wright | 08:52:00 | 2020-05-19 | | The Psychology of In | Leron Zinatullin | Peter Silverleaf | 02:12:00 | 2018-11-27 | | On Psychology - Illu | JZ Murdock | JZ Murdock | 01:49:00 | 2018-07-02 | | GCHQ - Centenary Edi | Richard Aldrich | Peter Noble | 25:48:00 | 2019-07-11 | | Project Hail Mary | Andy Weir | Ray Porter | 16:10:00 | 2021-05-04 | | Sid Meier's Memoir! | Sid Meier, Jennifer | Charles Constant | 08:32:00 | 2020-11-10 | | Docker in Action | Jeff Nickoloff | Aiden Humphreys | 10:12:00 | 2018-11-08 | | Cryptonomicon | Neal Stephenson | William Dufris | 42:44:00 | 2020-08-08 | | The Testament of Mar | Colm Tóibín | Meryl Streep | 03:06:00 | 2014-05-01 | | Anathem | Neal Stephenson | Oliver Wyman, Tavia | 32:25:00 | 2020-08-08 | | The Stranger in the | Michael Finkel | John Chancer | 06:08:00 | 2018-09-27 | | Xenos - Eisenhorn: W | Dan Abnett | Toby Longworth | 09:55:00 | 2017-09-27 | | Have Space Suit - Wi | Robert A. Heinlein | Mark Turetsky | 08:53:00 | 2014-02-11 | | Malleus - Eisenhorn: | Dan Abnett | Toby Longworth | 10:19:00 | 2017-09-27 | | Klara and the Sun | Kazuo Ishiguro | Sura Siu | 10:16:00 | 2021-03-02 | | Hereticus - Eisenhor | Dan Abnett | Toby Longworth | 09:48:00 | 2017-09-27 | | Ravenor - Warhammer | Dan Abnett | Toby Longworth | 11:50:00 | 2018-03-27 | | Sun and Steel | Yukio Mishima | Matthew Taylor | 02:36:00 | 2021-04-12 | | The Silver Ships - T | S. H. Jucha | Grover Gardner | 10:27:00 | 2015-06-30 | | Globe - Life in Shak | Catharine Arnold | Clare Staniforth | 09:22:00 | 2021-11-30 | | The Buried Giant | Kazuo Ishiguro | David Horovitch | 11:48:00 | 2015-03-03 | | Damned | Chuck Palahniuk | Sophie Amoss | 07:42:00 | 2021-10-12 | | Fallen Dragon | Peter F. Hamilton | John Lee | 26:30:00 | 2016-11-17 | | Escalation - Frontie | Ryk Brown | Jeffrey Kafer | 07:15:00 | 2020-02-11 | | The Revenant | Michael Punke | Jeff Harding | 09:54:00 | 2015-05-07 | | Rescue - Frontiers S | Ryk Brown | Jeffrey Kafer | 08:00:00 | 2016-12-06 | | Resurrection - Front | Ryk Brown | Jeffrey Kafer | 08:26:00 | 2017-04-18 | | Titus Groan - Gormen | Mervyn Peake | Saul Reichlin | 21:39:00 | 2014-06-19 | | Rebellion - Frontier | Ryk Brown | Jeffrey Kafer | 08:16:00 | 2017-08-01 | | The Edgar Allan Poe | Edgar Allan Poe | Jonathan Keeble, Pet | 59:59:00 | 2022-02-04 | | Ravenor Returned - W | Dan Abnett | Toby Longworth | 10:56:00 | 2018-03-27 | | Not Forever, but for | Chuck Palahniuk | Raphael Corkhill | 08:37:00 | 2023-09-05 | | Gormenghast | Mervyn Peake | Saul Reichlin | 22:56:00 | 2014-06-19 | | British Woodland - H | Ray Mears | Ray Mears | 09:22:00 | 2023-05-04 | | A Canticle for Leibo | Walter M. Miller Jr. | Tom Weiner | 10:55:00 | 2011-08-01 | | Weaving the Web - Th | Tim Berners-Lee | Tim Berners-Lee | 03:29:00 | 1999-12-16 | | Balance - Frontiers | Ryk Brown | Jeffrey Kafer | 10:53:00 | 2017-08-30 | | The Plum in the Gold | David Tod Roy - tran | George Backman | 17:35:00 | 2014-04-18 | | Stranger in a Strang | Robert A. Heinlein | Martin McDougall | 23:33:00 | 2012-12-06 | | Ravenor Rogue - Warh | Dan Abnett | Toby Longworth | 12:30:00 | 2018-01-31 | | Freakonomics - A Rog | Steven D. Levitt, St | Stephen J. Dubner | 07:50:00 | 2005-07-27 | | The Pragmatic Progra | David Thomas, Andrew | Anna Katarina | 09:55:00 | 2019-12-26 | | I'm Starting to Worr | Jason Pargin | Ari Fliakos | 12:44:00 | 2024-09-24 | | Birdsong | Sebastian Faulks | Harry Lloyd, Pippa B | 15:49:00 | 2023-06-15 | | The Luzhin Defense | Vladimir Nabokov | Mel Foster | 08:37:00 | 2010-12-20 | | The Three-Body Probl | Cixin Liu, Ken Liu | Daniel York Loh | 14:46:00 | 2023-02-23 | Provide feedback on this episode.
¿Estás seguro de que tu servidor Linux y tus contenedores Docker están a salvo de intrusos?
En este episodio exploramos las buenas prácticas oficiales de Docker y cómo aplicarlas en proyectos reales. Hablamos sobre qué evitar, qué optimizar y cómo escribir Dockerfile e imágenes más seguras, ligeras y eficientes. Un repaso directo, útil y práctico para cualquier desarrollador que quiera llevar sus contenedores al siguiente nivel.
At JupyterCon 2025, Jupyter Deploy was introduced as an open source command-line tool designed to make cloud-based Jupyter deployments quick and accessible for small teams, educators, and researchers who lack cloud engineering expertise. As described by AWS engineer Jonathan Guinegagne, these users often struggle in an “in-between” space—needing more computing power and collaboration features than a laptop offers, but without the resources for complex cloud setups. Jupyter Deploy simplifies this by orchestrating an entire encrypted stack—using Docker, Terraform, OAuth2, and Let's Encrypt—with minimal setup, removing the need to manually manage 15–20 cloud components. While it offers an easy on-ramp, Guinegagne notes that long-term use still requires some cloud understanding. Built by AWS's AI Open Source team but deliberately vendor-neutral, it uses a template-based approach, enabling community-contributed deployment recipes for any cloud. Led by Brian Granger, the project aims to join the official Jupyter ecosystem, with future plans including Kubernetes integration for enterprise scalability. Learn more from The New Stack about the latest in Jupyter AI development: Introduction to Jupyter Notebooks for DevelopersDisplay AI-Generated Images in a Jupyter Notebook Join our community of newsletter subscribers to stay on top of the news and at the top of your game. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
Register for FREE Infosec Webcasts, Anti-casts & Summits – https://poweredbybhis.com Chapters00:00 - PreShow Banter™ — Humans are Done03:04 - Louvre's video security password was ‘Louvre' – BHIS - Talkin' Bout [infosec] News 2025-11-1005:11 - Story # 1: I Tried the Robot That's Coming to Live With You. It's Still Part Human.15:14 - Story # 2: How to trade your $214,000 cybersecurity job for a jail cell25:14 - Story # 3: The Louvre's video security password was reportedly ‘Louvre'29:04 - Story # 4: Dangerous runC flaws could allow hackers to escape Docker containers32:58 - Story # 5: List of AI Tools Promoted by Threat Actors in Underground Forums and Their Capabilities40:00 - Story # 5b: GTIG AI Threat Tracker: Advances in Threat Actor Usage of AI Tools56:37 - BHIS Webcast – X-Typhoon - Not your Father's China with John Strand
En este episodio me sumerjo en uno de los entornos de escritorio más comentados y esperados del mundo open source: COSMIC.Mi podcast se centra en Linux y el software de código abierto, ofreciendo soluciones y métodos para mejorar la productividad, y al ver el hype alrededor de COSMIC, no pude resistirme. Lo instalé en mi ArchLinux con el objetivo de probarlo a fondo, ver cómo se comporta y, sobre todo, evaluar si realmente ofrece algo que me haga considerar migrar de mi combinación actual de GNOME y Niri.Y lo cierto es que me he encontrado con un escritorio interesante, que combina la opción por un Tiling Window Manager con la de un escritorio tradicional. Pero, como en toda herramienta, tengo mis peros.Análisis de COSMIC: Lo Bueno, Lo Malo y Mis Críticas SincerasLa Excelencia en Tiling: Sin lugar a dudas, lo que más me ha gustado de Cosmic es su gestor de ventanas tipo tiling. Sinceramente, es la mejor experiencia de Tiling Window Manager que he probado hasta la fecha en un entorno de escritorio tradicional. Esta es una gran ventaja sobre los gestores puros (i3, Sway, Niri) que te obligan a instalar y configurar un lanzador, un gestor de red, una barra de tareas, etc. Si buscas probar el tiling sin complicarte la vida, COSMIC es una gran opción.Productividad en Código: ¿Mejor que Niri? A pesar de la calidad de su Tiling, no alcanza el nivel de Niri para programar. Con Niri, tengo la facilidad de construir el espacio de trabajo que necesito en cada momento con una rapidez y sencillez inigualables. En este aspecto, COSMIC todavía no se acerca a la fluidez que busco.Los Auxiliares y Configuración: Cosmic incluye todas esas herramientas que completan la experiencia de usuario: barra de tareas, dock, notificaciones. La herramienta de configuración es bastante completa y muy al estilo GNOME, lo que la hace familiar para muchos usuarios. No tiene tantas opciones como GNOME, pero es funcional.La Decepción Estética (¡El Aspecto Visual de 2015!): Este es mi gran "pero". Desde mi punto de vista, el aspecto visual de Cosmic es propio de hace diez años. Es como volver al pasado. No está tan pulido como GNOME o KDE; la integración con herramientas de terceros es mejorable y, sinceramente, le queda un largo camino para ser un entorno de escritorio atractivo.Las Aplicaciones Nativas y mi Problema con Flatpak: Confieso que no me gustan las aplicaciones nativas de Cosmic; creo que el equipo debería centrarse en pulir el entorno para integrar perfectamente las aplicaciones de terceros. Pero lo que realmente "remata la fiesta" es la tienda de aplicaciones. Resulta que todas las aplicaciones que ofrece, o al menos las que vi, hay que instalarlas con Flatpak. Ya sabéis que no puedo con Flatpak; me parece una locura que para instalar una herramienta sencilla haya que descargar paquetes que ocupan una barbaridad, se integran pobremente y, a menudo, ni funcionan correctamente.Conclusión: El tiling de Cosmic funciona muy bien y es una gran puerta de entrada para los nuevos usuarios. Pero en el resto de aspectos, todavía tiene mucho que recorrer para estar a la altura de entornos más maduros como GNOME o KDE.Si buscas soluciones prácticas para la gestión de datos, la optimización de sistemas Linux o quieres ver la evolución de tecnologías clave como Docker, Neovim, Rust o Traefik, este episodio te dará una perspectiva útil sobre el futuro de los escritorios Linux.Más información y enlaces en las notas del episodio
runC flaws could allow hackers to escape Docker containers Lost iPhone scam warning Landfall Android spyware targets Samsung Galaxy phones Huge thanks to our sponsor, Vanta What's your 2 AM security worry? Is it "Do I have the right controls in place?" Or "Are my vendors secure?" ....or the really scary one: "how do I get out from under these old tools and manual processes? Enter Vanta. Vanta automates manual work, so you can stop sweating over spreadsheets, chasing audit evidence, and filling out endless questionnaires. Their trust management platform continuously monitors your systems, centralizes your data, and simplifies your security at scale. Vanta also fits right into your workflows, using AI to streamline evidence collection, flag risks, and keep your program audit-ready—ALL…THE…TIME. With Vanta, you get everything you need to move faster, scale confidently—and get back to sleep. Get started at vanta.com/headlines Find the stories behind the headlines at CISOseries.com.
Today we have Eric Curtin from the Docker Model Runner project on the show the Docker projects attempt at bringing the ease of Docker to running AI models on your local hardware or deployed as a service.==========Support The Channel==========► Patreon: https://www.patreon.com/brodierobertson► Paypal: https://www.paypal.me/BrodieRobertsonVideo► Amazon USA: https://amzn.to/3d5gykF► Other Methods: https://cointr.ee/brodierobertson==========Guest Links==========Docker Site: https://www.docker.com/Announcement: https://www.docker.com/blog/introducing-docker-model-runner/Docs: https://docs.docker.com/ai/model-runner/==========Support The Show==========► Patreon: https://www.patreon.com/brodierobertson► Paypal: https://www.paypal.me/BrodieRobertsonVideo► Amazon USA: https://amzn.to/3d5gykF► Other Methods: https://cointr.ee/brodierobertson=========Video Platforms==========
In this episode of the Ardan Labs Podcast, Bill Kennedy talks with Salah Mahmud, a medical researcher specializing in epidemiology and medical statistics. Salah shares his journey from growing up in Libya under Gaddafi's regime to conducting cutting-edge research in Canada on the connection between influenza and heart attacks. He discusses the challenges of running large-scale observational studies, the bureaucratic barriers to accessing medical data, and the importance of diversity in health research. Salah also reflects on his early entrepreneurial ventures, his discovery of programming during medical school, and how resilience and adaptability shaped his personal and professional journey.00:00 Introduction02:03 Research on Influenza and Heart Attacks05:53 Challenges in Data Access16:51 Life in Libya Under Gaddafi21:32 From Medicine to Programming41:18 WHO Collaboration and Education Abroad57:13 Disappearance and New Beginnings01:09:33 Immigration and Adaptation in Canada01:15:45 Balancing Medicine and Technology01:21:22 Family, Culture, and Reflection01:25:37 Current Research and Future GoalsConnect with Salah: Email: salah.mahmud@umanitoba.caLinkedIn: https://www.linkedin.com/in/salah-mahmud-4177285a/Mentioned in this Episode:Golang: https://go.dev/Want more from Ardan Labs? You can learn Go, Kubernetes, Docker & more through our video training, live events, or through our blog!Online Courses : https://ardanlabs.com/education/ Live Events : https://www.ardanlabs.com/live-training-events/ Blog : https://www.ardanlabs.com/blog Github : https://github.com/ardanlabs
Si confías tus recuerdos a Immich (tu servidor de fotos self-hosted), la seguridad de esos datos no es opcional, es obligatoria. En este episodio, te muestro el método definitivo para garantizar la seguridad de tus fotos y metadatos sin depender de soluciones comerciales.El desafío de Immich es realizar un backup coherente que sincronice los archivos y la base de datos (PostgreSQL) al mismo tiempo. Para resolver esto, he creado una solución robusta y práctica.Veremos en detalle:El Orquestador: Te presento mi proyecto rubadb, la herramienta que he desarrollado para automatizar el flujo de backup de principio a fin, incluyendo la gestión de retención automática.El Especialista en Bases de Datos: Analizamos a fondo postgresus, la utilidad que nos asegura un dump limpio y comprimido de la base de datos de PostgreSQL (la clave de tus metadatos), con soporte para múltiples destinos (S3, Dropbox, etc.).La Configuración Práctica: Te explico cómo integrar ambas herramientas en un entorno Docker para conseguir un proceso automatizado, ultra-seguro y que te dará la tranquilidad de saber que tus recuerdos están a salvo.Si quieres llevar la protección de tus aplicaciones self-hosted al siguiente nivel con soluciones de código abierto y prácticas, este episodio es para ti.Más información y enlaces en las notas del episodio
SANS Internet Stormcenter Daily Network/Cyber Security and Information Security Stormcast
How to Collect Memory-Only Filesystems on Linux Systems Getting forensically sound copies of memory-only file systems on Linux can be tricky, as tools like dd do not work. https://isc.sans.edu/diary/How%20to%20collect%20memory-only%20filesystems%20on%20Linux%20systems/32432 Microsoft Azure Front Door Outage Today, Microsoft s Azure Front Door service failed, leading to users not being able to authenticate to various Azure-related services. https://azure.status.microsoft/en-us/status Docker-Compose Vulnerability A vulnerability in docker-compose may be used to trick users into creating files outside the docker-compose directory https://github.com/docker/compose/security/advisories/GHSA-gv8h-7v7w-r22q
North Melbourne forward Kate Shierlaw joins the W Show to discuss the Kangaroos’ record-breaking run, Gold Coast’s trade plans, and the Fremantle star who could be looking for a new homeSee omnystudio.com/listener for privacy information.
In this episode, we range from ice-cold mornings and sunny Colorado skies to a deep dive on home mining, heat reuse, open hardware, and sovereign home automation. We recap getting featured in Forbes on Heat Punk projects and how mainstream coverage is finally grokking mining-as-heat, Canon's heating-first designs, and Bitmain's market dominance risks. We share real-world progress: integrating Canaan home miners with Home Assistant via APIs and Node-RED, using Zigbee sensors for room-aware thermostatic control, solar and TOU-aware automations, and the vision for a sovereign “miner control hub” box built on Raspberry Pi 5. We get nerdy on RISC‑V vs ARM, open firmware, and the Libre Board + Mujina roadmap, with detours through customs-destroyed SMD parts, packaging HydroPool for Docker, and the power of public, self-hosted pools after a solo-Block win with a NerdQAX. We also cover privacy and the surveillance creep: doorbells, cars, app signing, and why self-hosted tools (Pi-hole, PFsense, Mullvad, Signal, Proton/Tutanota) matter. We discuss HPC pivots by large miners, grid vs. heat-reuse economics, Canaan's momentum in home heating, and the imminent Telehash on HydroPool with StartOS packaging on deck. Plus, the Stealth Miner enclosure, Bitaxe-powered heat projects, and shoutouts to the open-source crew making sovereignty practical at home, one sensor, miner, and Docker container at a time.
En este episodio, me enfrento a un desafío de rendimiento real: el consumo de CPU de PostgreSQL se dispara, pero sin alta actividad de lectura/escritura.Viajaremos a través de un diagnóstico detallado utilizando herramientas nativas de Postgres como pg_stat_activity y pg_stat_statements para desenmascarar las causas ocultas:La Sobrecarga de Conexión: Descubriremos cómo un simple healthcheck de Docker (pg_isready) configurado incorrectamente puede paralizar tu servidor por el alto overhead de gestión de procesos.El Cuello de Botella de la Aplicación: Analizamos y corregimos un error de diseño de código en Rust/Axum donde se recompilan Expresiones Regulares (Regex) en cada petición, consumiendo innecesariamente ciclos de CPU.Una lección práctica esencial para cualquiera que gestione bases de datos, despliegues en Docker o desarrolle aplicaciones eficientes en Rust en entornos Linux. Aprende a identificar estos fallos y a optimizar tu código mediante la precompilación de Regex utilizando TryFrom y Arc.Más información y enlaces en las notas del episodio
For memberships: join this channel as a member here:https://www.youtube.com/channel/UC_mGuY4g0mggeUGM6V1osdA/joinSummaryIn this conversation, Ian discusses the evolution and significance of Unikernels and NanoVMs, emphasizing their potential to enhance security and performance in cloud computing. He explains the historical context of operating systems, the limitations of traditional systems, and how Unikernels offer a streamlined alternative. Ian also highlights the unique features of NanoVMs, their integration capabilities, and the challenges faced in the ecosystem. The discussion concludes with insights into the future of Unikernels and the ongoing developments in the field.takeaways.Unikernels are a specialized type of operating system designed for cloud environments.The evolution of operating systems has led to the need for more efficient solutions like Unikernels.Unikernels can significantly reduce security vulnerabilities compared to traditional systems.NanoVMs provide a unique approach to Unikernels with a focus on performance and security.Integrations with existing tools and libraries are crucial for the adoption of Unikernels.The ecosystem around Unikernels is still developing, with many opportunities for growth.Unikernels eliminate the need for complex orchestration and management layers.The future of Unikernels includes tighter integrations with cloud services and improved developer experiences.Security features in Unikernels are designed to address modern threats effectively.The potential for Unikernels to transform application deployment is significant, with many untapped possibilities.Chapters00:00 Introduction to Unikernels and NanoVMs04:24 The Evolution of Operating Systems11:24 Understanding Unikernels vs. Traditional Systems17:20 Security Implications of Unikernels26:17 NanoVMs: Architecture and Unique Features38:44 Security Concerns in Unikernels41:05 Integration and Support for GPUs44:02 Cloud Support and Deployment45:51 Avoiding Bloat in Integrations51:54 Developer's Perspective on Unikernels59:18 Limitations and Future of UnikernelsImportant Links:https://ops.cityhttps://nanos.orghttps://repo.ops.cityhttps://nanovms.com/dev/tutorialsFor memberships: join this channel as a member here:https://www.youtube.com/channel/UC_mGuY4g0mggeUGM6V1osdA/joinDon't forget to like, share, and subscribe for more insights!=============================================================================Like building stuff? Try out CodeCrafters and build amazing real world systems like Redis, Kafka, Sqlite. Use the link below to signup and get 40% off on paid subscription.https://app.codecrafters.io/join?via=geeknarrator=============================================================================Database internals series: https://youtu.be/yV_Zp0Mi3xsPopular playlists:Realtime streaming systems: https://www.youtube.com/playlist?list=PLL7QpTxsA4se-mAKKoVOs3VcaP71X_LA-Software Engineering: https://www.youtube.com/playlist?list=PLL7QpTxsA4sf6By03bot5BhKoMgxDUU17Distributed systems and databases: https://www.youtube.com/playlist?list=PLL7QpTxsA4sfLDUnjBJXJGFhhz94jDd_dModern databases: https://www.youtube.com/playlist?list=PLL7QpTxsA4scSeZAsCUXijtnfW5ARlrsNStay Curios! Keep Learning!
En este episodio de "atareao con Linux", abordamos una frustración común: la sobrecarga de complejidad en el mundo del blogging. Si has intentado usar WordPress y te has cansado de gestionar plugins, temas y vulnerabilidades, o si las soluciones de Static Site Generator (SSG) te parecen excesivas para simplemente publicar notas y código, Noet es la solución que has estado buscando.Noet es una plataforma de blogging de código abierto con una filosofía clara: priorizar la escritura. Su diseño se basa en quitar todo lo que se interpone entre tú y la publicación de tu contenido. Es, esencialmente, un editor de texto avanzado que guarda posts en una base de datos y los sirve como un sitio web limpio y legible.La verdadera magia de Noet reside en su simplicidad técnica, lo cual lo hace perfecto para nuestro entorno Linux (VPS, Raspberry Pi, o tu servidor local):Single Binary (Go): Todo el backend se compila en un único ejecutable (escrito en Go), lo que facilita enormemente el despliegue y el mantenimiento en cualquier plataforma Linux.SQLite para la Gestión de Datos: En lugar de depender de bases de datos externas como MySQL o PostgreSQL, Noet usa SQLite. Esto significa que todos tus posts y configuraciones se almacenan en un solo archivo, noet.db. Esta característica es fundamental para una gestión de datos eficiente y para realizar copias de seguridad de forma increíblemente sencilla.Despliegue con Docker: Fieles a nuestro estilo práctico, te mostramos el archivo docker-compose.yaml necesario para poner Noet en marcha en cuestión de minutos. Si ya usas Docker para servicios como Traefik, Syncthing o tus bases de datos [cite: 2025-07-15], añadir Noet a tu stack es trivial.Para el escritor técnico o el power user de Linux, Noet brilla en su editor:Soporte Markdown Nativo: Usa la sintaxis que ya conoces.Código y LaTeX: El editor soporta resaltado de sintaxis para bloques de código y permite incrustar ecuaciones matemáticas con LaTeX/KaTeX. Es ideal para documentar tus proyectos o publicar tutoriales avanzados.Auto-guardado: No pierdas ni una línea de lo que escribes.Sencillez en Imágenes: Arrastra y suelta para subir imágenes y gestiona su tamaño con un clic.Si buscas mejorar tu productividad, simplificar tu infraestructura y tener un blog que se sienta tan ligero y moderno como Neovim u Obsidian [cite: 2025-07-15] pero listo para publicar en la web, tienes que probar Noet.Escucha el episodio para obtener todos los comandos, el archivo docker-compose y los mejores consejos de uso.Más información y enlaces en las notas del episodio
This is a recap of the top 10 posts on Hacker News on October 22, 2025. This podcast was generated by wondercraft.ai (00:30): MinIO stops distributing free Docker imagesOriginal post: https://news.ycombinator.com/item?id=45665452&utm_source=wondercraft_ai(01:52): Scripts I wrote that I use all the timeOriginal post: https://news.ycombinator.com/item?id=45670052&utm_source=wondercraft_ai(03:14): Greg Newby, CEO of Project Gutenberg Literary Archive Foundation, has diedOriginal post: https://news.ycombinator.com/item?id=45666510&utm_source=wondercraft_ai(04:36): Internet's biggest annoyance: Cookie laws should target browsers, not websitesOriginal post: https://news.ycombinator.com/item?id=45667866&utm_source=wondercraft_ai(05:58): Google flags Immich sites as dangerousOriginal post: https://news.ycombinator.com/item?id=45675015&utm_source=wondercraft_ai(07:20): Meta is axing 600 roles across its AI divisionOriginal post: https://news.ycombinator.com/item?id=45671778&utm_source=wondercraft_ai(08:42): Willow quantum chip demonstrates verifiable quantum advantage on hardwareOriginal post: https://news.ycombinator.com/item?id=45670443&utm_source=wondercraft_ai(10:04): AI assistants misrepresent news content 45% of the timeOriginal post: https://news.ycombinator.com/item?id=45668990&utm_source=wondercraft_ai(11:26): Greenland's national telco, Tusass, signs new agreement with EutelsatOriginal post: https://news.ycombinator.com/item?id=45665796&utm_source=wondercraft_ai(12:48): French ex-president Sarkozy begins jail sentenceOriginal post: https://news.ycombinator.com/item?id=45665311&utm_source=wondercraft_aiThis is a third-party project, independent from HN and YC. Text and audio generated using AI, by wondercraft.ai. Create your own studio quality podcast with text as the only input in seconds at app.wondercraft.ai. Issues or feedback? We'd love to hear from you: team@wondercraft.ai
Following the success of its DXP series, UGREEN has announced the launch of its new DH Series, designed to make network-attached storage (NAS) more accessible for entry-level users and those with essential data storage needs. The lineup includes two models, NASync DH2300 and NASync DH4300 Plus, with the DH2300 officially available starting October 15. NASync DH2300: Accessible NAS for Everyone The NASync DH2300 is the ideal first step into personal NAS, tailored for cloud drive and hard drive users, as well as home entertainment enthusiasts seeking a secure and more efficient way to manage growing data. With a 2-bay SATA configuration supporting up to 60TB (30TB per drive), it effortlessly handles vast libraries of 4K videos, high-resolution photos, and large documents, without relying on third-party cloud services. Running on UGOS Pro, UGREEN's intuitive operating system, the device offers a guided setup process that allows even first-time NAS users to complete installation in under ten minutes. The all-in-one UGREEN NAS app integrates file management, automatic photo backup, and media streaming to TVs through a single interface, removing the need for multiple apps. NFC quick connection further simplifies access, allowing users to connect a smartphone with just a tap. Security is built into every layer. DH2300 ensures full control of personal data through local storage, protected by TLS/SSL, RSA, and AES encryption, two-factor authentication, and certifications from TÜV and TRUSTe. The built-in Security Manager adds 24/7 threat monitoring and scheduled virus scans for complete peace of mind. With additional hardware features including a 1GbE LAN port for stable connectivity, 4K 60Hz HDMI output, and multiple RAID modes for flexible performance and redundancy, the DH2300 delivers simplicity, security, and versatility. It is the go-to NAS for users moving from cloud-based to private local storage for the first time. NASync DH4300 Plus: Designed for Growth and Collaboration For users who need more storage and performance, NASync DH4300 Plus is a powerful choice. Its 4-bay SATA setup supports up to 120TB, ideal for home offices, creative teams, and media-heavy workflows. With stable multitasking across data management, streaming, and collaboration, it features a 2.5GbE LAN port delivering theoretical speeds of up to 312.5MB/s, and supports RAID 5, 6, and 10 for enhanced redundancy and data protection. USB-A and USB-C 3.2 ports offer fast connectivity, while Docker support enables flexible deployment. DH4300 Plus combines professional-grade capabilities with an intuitive interface and the same robust security foundation as DH2300. The new DH Series represents UGREEN's ongoing commitment to creating NAS solutions that truly fit users' lifestyles. By simplifying setups, streamlining daily workflows, and strengthening data protection, the DH Series makes intelligent storage a reality for beginners, families, and small teams alike. DH2300 is now available at €209.99 in the EU and £169.99 in the UK. DH4300 Plus is currently offered at €429.99 in the EU and £359.99 in the UK. For more details, check the UGREEN website. More about Irish Tech News Irish Tech News are Ireland's No. 1 Online Tech Publication and often Ireland's No.1 Tech Podcast too. You can find hundreds of fantastic previous episodes and subscribe using whatever platform you like via our Anchor.fm page here: https://anchor.fm/irish-tech-news If you'd like to be featured in an upcoming Podcast email us at Simon@IrishTechNews.ie now to discuss. Irish Tech News have a range of services available to help promote your business. Why not drop us a line at Info@IrishTechNews.ie now to find out more about how we can help you reach our audience. You can also find and follow us on Twitter, LinkedIn, Facebook, Instagram, TikTok and Snapchat.
Stories come to life in Fort Worth! That was the underlying theme of this years State of the City event. Fort Worth Roots was honored to attend and get a chance to record with some of Cowtown's best. Up first we have Allison Docker with Fort Worth Green Space Champion. Next, Fort Worth City Manager Jay Chapa. Our third recording is with Kelly Porter, Assistant Director of Regional Transportation Planning & Innovation. To close the episode out and talk about some of the things she covered in her speech is Mayor Mattie Parker! Special thanks to everyone that made space for Fort Worth Roots so that we could be part of this special day. Follow along to see what else the Fort Worth Chamber is up to on Instagram @ftwchamber Find information about the city of Fort Worth online at: www.fortworthtexas.gov See the full video of the event with this link : https://youtu.be/rBczPI0nszk?si=w6I894kdjnPaN3D_
In this episode of the Ardan Labs Podcast, Bill Kennedy talks with Zack Holland, CEO & Founder of Averi AI, about his journey from early life in Ecuador's Amazon rainforest to building an AI-powered marketing platform. Zack shares lessons from early business ventures, the challenges of running startups, and the evolution of his entrepreneurial mindset. They explore how Averi AI helps marketers become more creative and efficient, the importance of data security and trust in AI, and what it takes to innovate in a rapidly evolving digital landscape.00:00 Introduction02:54 Marketing and AI Evolution05:58 Changing Digital Landscape09:04 Early Life and Influences12:02 From Ecuador to Utah18:00 First Business in High School29:53 LLMs and Entrepreneurship39:44 Lessons from Failure43:46 Starting a Marketing Agency56:19 Founding Avery AI01:08:10 Trust and Data Security01:13:14 Marketing and AI Adoption01:17:58 AI Challenges and Opportunities01:26:00 Contact Info Connect with Zack: Linkedin: https://www.linkedin.com/in/zackhollandX: https://x.com/zack_hollandMentioned in this Episode:Averi AI: https://www.averi.ai/Want more from Ardan Labs? You can learn Go, Kubernetes, Docker & more through our video training, live events, or through our blog!Online Courses : https://ardanlabs.com/education/ Live Events : https://www.ardanlabs.com/live-training-events/ Blog : https://www.ardanlabs.com/blog Github : https://github.com/ardanlabs
Varios oyentes de este podcast me habéis dicho alguna vez que no usáis Docker, que preferís virtualizar los servicios que desplegáis en vuestra casa y que yo tenía que probar Proxmox. Pues bien, lo he hecho y cuento aquí mis sensaciones iniciales.
Talk Python To Me - Python conversations for passionate developers
Python in 2025 is different. Threads really are about to run in parallel, installs finish before your coffee cools, and containers are the default. In this episode, we count down 38 things to learn this year: free-threaded CPython, uv for packaging, Docker and Compose, Kubernetes with Tilt, DuckDB and Arrow, PyScript at the edge, plus MCP for sane AI workflows. Expect practical wins and migration paths. No buzzword bingo, just what pays off in real apps. Join me along with Peter Wang and Calvin Hendrix-Parker for a fun, fast-moving conversation. Episode sponsors Seer: AI Debugging, Code TALKPYTHON Agntcy Talk Python Courses Links from the show Calvin Hendryx-Parker: github.com/calvinhp Peter on BSky: @wang.social Free-Threaded Wheels: hugovk.github.io Tilt: tilt.dev The Five Demons of Python Packaging That Fuel Our ...: youtube.com Talos Linux: talos.dev Docker: Accelerated Container Application Development: docker.com Scaf - Six Feet Up: sixfeetup.com BeeWare: beeware.org PyScript: pyscript.net Cursor: The best way to code with AI: cursor.com Cline - AI Coding, Open Source and Uncompromised: cline.bot Watch this episode on YouTube: youtube.com Episode #524 deep-dive: talkpython.fm/524 Episode transcripts: talkpython.fm Theme Song: Developer Rap
¿Tu Watchtower te ha dejado alguna vez un servicio crítico caído? Es hora de automatizar la seguridad de tus contenedores Docker, ¡pero con control total y una Interfaz Gráfica (Web UI)! Tugtainer es la alternativa que estabas buscando para decirle adiós a las vulnerabilidades y a los "desastres del sábado". Si gestionas tu propio stack en Linux, esta herramienta self-hosted te va a cambiar la vida. Escucha y descubre cómo tener contenedores siempre al día, pero con seguridad.Hay dos cosas que obsesionan a cualquier administrador de sistemas que utiliza Docker en entornos self-hosted: las copias de seguridad de las bases de datos y la actualización constante de las imágenes para evitar vulnerabilidades. Aunque la actualización automática es fundamental como acción preventiva, si se hace de forma completamente desatendida, puede causar más de un trastorno.Durante años, he usado Watchtower para la mayoría de mis servicios. Sin embargo, esta herramienta, aunque se integra perfectamente con Docker y las etiquetas, tiene dos grandes problemas: carece de una interfaz gráfica para ver qué está ocurriendo y lleva tiempo sin recibir actualizaciones.El Dilema del Control:Los servicios críticos, como las páginas web que administro (con stacks de WordPress, MariaDB y Nginx), no pueden permitirse caídas. Por eso, dejé la política de actualizaciones diarias y la cambié por una revisión semanal (los sábados). Hoy, vamos a resolver este dilema: ¿Cómo conseguimos la automatización de la seguridad sin sacrificar la estabilidad?Llega Tugtainer: El Control Gráfico que NecesitabasEn este episodio, te presento una herramienta nueva y prometedora (¡con solo un mes de vida!) que se posiciona como una alternativa a Watchtower y Ouroboros. Se trata de Tugtainer, la solución que añade una Web UI completa a la gestión de actualizaciones de Docker.Lo que Aprenderás en el Episodio:Por qué mi stack web (con dependencias service_healthy) sigue dándome problemas al actualizar, y la lección aprendida.Las advertencias cruciales del desarrollador de Tugtainer: por qué no se recomienda para entornos de producción (¡al menos por ahora!).Análisis a fondo de las siete características de Tugtainer que te dan control total:Configuración por Contenedor: Decidir si un servicio CRÍTICO (como Traefik) solo se verifica o si se auto-actualiza.Programación Crontab: Control total sobre cuándo se lanzan las comprobaciones.Autenticación y Notificaciones: Seguridad y visibilidad al instante.Limpieza de imágenes: Adiós a las imágenes obsoletas que ocupan espacio.Mi propia implementación de Tugtainer con Docker Compose, Traefik y Dockge (¡una herramienta que deberías conocer!).Si utilizas Linux, Docker y buscas maximizar tu productividad y seguridad en tu VPS o Raspberry Pi, este episodio es una guía esencial para pasar de la automatización ciega a la automatización inteligente.¡Dale al play y descubre si Tugtainer se queda o no en mi propio stack de atareao!Soy Lorenzo Carbonell, "atareao". En este podcast me centro en el software libre y Linux. Mi estilo es práctico y te traigo soluciones, métodos y tutoriales para mejorar la productividad, gestionar datos y optimizar sistemas Linux. Si te interesa Docker, Neovim, Rust, Syncthing o configurar servicios en plataformas como Raspberry Pi o VPS, ¡suscríbete!Más información y enlaces en las notas del episodio
Foundations of Amateur Radio The other day I received an email from Frank K4FMH asking me about an idea I'd worked on some time ago, namely the notion that I might monitor solar flux at home using a software defined radio. At the time I was attempting to get some software running on my PlutoSDR and got nowhere fast. Before I continue, a PlutoSDR, or more formally an ADALM Pluto Active Learning Module by Analog Devices, is both a computer and a software defined radio receiver and transmitter in a cute little blue box. I've talked about this device before. It's an open design, which means that both the software and hardware are documented and available straight from the manufacturer. Out of the box it covers 325 MHz to 3.8 GHz. You can connect to a PlutoSDR using USB or via the network, wireless or Ethernet, though I will mention that neither of those last two is currently working for me, but more on that later. Encouraged by Frank's email, I set out to explore further and came across a 2019 European GNU Radio days workshop, which discussed some of the tools that are available for the PlutoSDR, accompanied by two PDF documents walking you through the experience. One comment around why the PlutoSDR uses networking as one of the connectivity options spoke to me. From a usability perspective, networking makes it easier to access the PlutoSDR from a virtual machine, since most of the time that already has network connectivity, whereas USB often requires drivers. As you might recall, network connectivity is one of the many things that I'm trying to achieve with a project that I'm calling Bald Yak, since by the time we're done, there's not going to be much hair left from all the Yak Shaving. The Bald Yak project aims to create a modular, bidirectional and distributed signal processing and control system that leverages GNU Radio. As a result, I set about trying to actually walk myself through those PDF tutorials .. and got stuck on the first sentence on the first page, which helpfully states: "The necessary prerequisites have been installed on the local lab machine." It went on to supply a link to a page with instructions on how to acquire those very same prerequisites. Two days later, after much trial and error, I can now report that I too have these installed and because I cannot help myself, I made it into a Docker container and published this on my VK6FLAB GitHub page. To put it mildly, there's a few moving parts and plenty of gotchas. As an aside, if you think that installing Docker is harder than installing these tools, I have some news for you .. trust me .. by a long shot .. it's not. Right now I'm working on writing the documentation that accompanies this project such that you can actually use it without needing to bang your head against the desk in frustration. Mind you, the documentation part of this is non-trivial. For reasons I don't yet understand, my Pluto does not want to talk to the network directly over either WiFi or Ethernet, and connecting over USB through a virtual machine inside a Docker container is giving me headaches, so right now I'm connected across the network to a Raspberry Pi that's physically connected to the Pluto. As a result, I can now use the tools inside my Docker container, connected to the Pluto through the Pi and if you're curious, 'iiod' is the tool to make that happen .. more documentation. At this point you might well ask, why bother? This is a fair question. Let me see if I can give you an answer that will satisfy. Monitoring solar flux typically occurs at 2.8 GHz, which is outside the range of RTL-SDR dongles which top out at about 1.7 GHz. For the PlutoSDR however, it's almost perfectly within the standard frequency range. One of the tools that is introduced by the talk is an application called 'iio-scope', which as the name suggests, is an oscilloscope for 'iio' or Industrial I/O devices, of which the PlutoSDR is one. As an aside, the accelerometer in your laptop, the battery voltage, the CPU temperatures, fans, and plenty of others, are all 'iio' devices that you can look at with various tools. So, once I've finished the tutorials, I suspect that I will understand a little better how some of the various parts of the PlutoSDR hang together, and I can set it up to monitor 2.8 GHz. Of course, that's only step one, the next step is to make a Raspberry Pi record the power levels over time, better still, record it on the PlutoSDR itself, and see if we can actually notice any change .. without requiring anything fancy like a special antenna, some massive filters, a special mount and all the other fun and games that no doubt will reveal themselves in good time. It also means that, if I got this right, I have the beginnings of the bits needed to get the PlutoSDR to talk to GNU Radio. Why? Because I can, and because Frank asked, also Yak Shaving. I'm Onno VK6FLAB
Nassim Eddequiouaq is co-founder and CEO of Bastion, a pioneer in regulated stablecoin infrastructure and NYDFS-certified provider. Bastion is the stablecoin issuance platform for financial institutions and enterprises. Prior to founding Bastion, Nass was the Chief Information Security Officer at a16z Crypto, and held senior management roles across Security and Infrastructure at Facebook, Anchorage, Docker, and Apple. He received a M.S. in Computer Science from Ecole d'Ingénieurs en Informatique. In this conversation, we discuss:- What happened on the 10/10 crypto crash? - Winners and losers after the crypto crash - Bridging traditional finance and digital assets through enterprise-ready solutions - The diverse use cases of stablecoins - Why stablecoins (especially USD-pegged) are poised for mass enterprise adoption - The growing interest in branded stablecoins - Bastion's NYDFS trust charter - GENIUS Act and STABLE Act - Why regulatory clarity is critical - Privacy for stablecoin users BastionX: @BastionPlatformWebsite: bastion.comLinkedIn: BastionNassim EddequiouaqX: @nassyweazyLinkedIn: Nassim Eddequiouaq---------------------------------------------------------------------------------This episode is brought to you by PrimeXBT.PrimeXBT offers a robust trading system for both beginners and professional traders that demand highly reliable market data and performance. Traders of all experience levels can easily design and customize layouts and widgets to best fit their trading style. PrimeXBT is always offering innovative products and professional trading conditions to all customers. PrimeXBT is running an exclusive promotion for listeners of the podcast. After making your first deposit, 50% of that first deposit will be credited to your account as a bonus that can be used as additional collateral to open positions. Code: CRYPTONEWS50 This promotion is available for a month after activation. Click the link below: PrimeXBT x CRYPTONEWS50
¡Atención usuarios de Arch Linux! En este nuevo episodio de Atareao con Linux, te desvelo la herramienta TUI (Terminal User Interface) que está cambiando por completo la forma de gestionar paquetes, tanto de los repositorios oficiales como del vasto ecosistema del Arch User Repository (AUR). Si el factor productividad es más importante para ti que el factor ilusión de las nuevas versiones, y valoras la filosofía KISS (Keep It Simple, Stupid) y el modelo rolling release, este tutorial práctico es crucial para optimizar tu flujo de trabajo.Mi nombre es Lorenzo Carbonell ("atareao"), y mi podcast se centra en ofrecer soluciones para cualquier cosa que quieras hacer con Linux. Este episodio es el ejemplo perfecto: te muestro cómo simplificar lo que antes requería múltiples comandos.Comenzamos profundizando en la que es, para muchos, la razón principal para elegir Arch Linux: el AUR (Arch User Repository). Este es un gran repositorio comunitario que te da acceso a una cantidad inmensa de software, incluyendo herramientas de nicho, versiones *-git y *-svn con las últimas características de desarrollo.Pero, ¿cómo gestionamos este poder?El AUR no contiene binarios, sino PKGBUILDs—scripts de compilación escritos por la comunidad que instruyen a tu sistema a descargar, verificar, compilar e instalar el paquete. Para manejar esto, necesitas un AUR Helper.He estado utilizando Paru durante mucho tiempo, y sigo encantado con él. Es una herramienta escrita en Rust que combina las funciones del gestor oficial pacman y la gestión de AUR.Características clave de Paru para la productividad:Herramienta Unificada.Seguridad y Transparencia.Gestión de Huérfanos.Y ahora, la estrella del episodio: Pacsea. Esta es una nueva herramienta de terminal, presentada como una TUI, que simplifica enormemente las tareas de búsqueda e instalación. También escrita en Rust, pacsea ofrece una interfaz intuitiva para buscar, filtrar e instalar paquetes.Por qué Pacsea es un game changer para tu productividad:Diseño de Tres Paneles: Interfaz clara con resultados, búsquedas recientes/instalación e información detallada del paquete.Seguridad Visual: Incorpora un visor de PKGBUILD que puedes abrir con Ctrl+x o un solo clic. Esto te permite revisar el código antes de la compilación e instalación.Instalación por Cola (Queueing): Pulsa la tecla Espacio para añadir varios paquetes a la cola y confirmar la instalación por lotes con Enter.Búsqueda Instantánea y Optimizada: El sistema de búsqueda utiliza debouncing para ser inmediato y simultáneo en repos oficiales y AUR.Filtros Clicables: Puedes filtrar los resultados directamente haciendo clic en las etiquetas [AUR], [core], [extra], etc..Información Esencial al Instante: Muestra la popularidad del AUR (estrellas), el estado de los servicios Arch/AUR y las noticias críticas que requieren intervención manual.Aunque siempre se puede exprimir más a herramientas nativas como paru y pacman, la facilidad de uso y la eficiencia de Pacsea la convierten en una herramienta imprescindible para mi flujo de trabajo en Arch Linux.Si buscas soluciones claras y directas para mejorar tu productividad, gestionar servidores web, proxies inversos (como Traefik) o utilizar herramientas avanzadas (como Docker, Neovim o Rust) [cite: 2025-07-15], este podcast es tu guía.
Si tienes un servidor Linux expuesto a Internet, ya sea un VPS o una Raspberry Pi alojando tus servicios Docker, este es un episodio que no te puedes saltar. Detrás de ese proxy inverso (Traefik es mi elección), se esconde un tráfico que rara vez revisamos, y te aseguro que no todo el mundo tiene buenas intenciones.Tras un incidente reciente que me obligó a abrir mi servidor al mundo (y no solo a España, como lo tenía restringido inicialmente), la cantidad de visitantes desconocidos y peticiones curiosas que encontré me hizo poner manos a la obra. No es solo un tema de seguridad; es de recursos.Cada visita cuesta. Sí, has oído bien. Cada interacción con tu servidor requiere un gasto de CPU y memoria RAM. Los bots y scanners que buscan vulnerabilidades o hacen peticiones inútiles están consumiendo silenciosamente la capacidad de tu sistema, dejando menos para tus visitas de calidad (las que realmente quieres). Es esencial saber quién te visita, dónde va, y con qué intenciones, para poder actuar y liberar esos recursos.Mi objetivo, como siempre en atareao con Linux, era encontrar una solución de código abierto que fuera sencilla de implementar y, crucialmente, que no se llevara por delante todos los recursos de mi propio servidor.El punto de partida de la investigación es siempre el access.log de Traefik, que es el registro fundamental de todas las peticiones.Estuve probando distintas combinaciones, incluyendo algunas pesadas y complejas, como:Vector, Prometheus, Grafana y Loki.Vector, Victorialogs, Grafana y Loki.Si bien estas son soluciones potentes, su complejidad y el alto consumo de recursos me hicieron descartarlas. La solución no debe ser un problema de rendimiento en sí misma.Finalmente, di con la combinación que es simple, eficiente y con la que estoy realmente enamorado por su facilidad de uso e implementación.Vector es la herramienta clave para recopilar, transformar y enrutar todos tus logs, métricas y trazas. Es de código abierto, hasta 10 veces más rápido que cualquier alternativa y es lo que me permite un enriquecimiento de datos sin precedentes.En este episodio aprenderás cómo:Configurar el compose.yml de Vector en tu entorno Docker.Utilizar las Transforms de Vector para parsear los logs de Traefik.Integrar la base de datos GeoIP (GeoLite2-City.mmdb) para geolocalizar la IP de procedencia de cada petición.Enrutar los logs enriquecidos a la base de datos de destino.OpenObserve (O2) es la plataforma de observabilidad nativa de la nube que unifica logs, métricas y trazas en una única interfaz. Es la alternativa que he adoptado a soluciones como ElasticSearch y se ha convertido en una herramienta imprescindible en mi día a día.Es increíblemente sencillo de instalar y configurar (lo tienes funcionando en minutos).Es el lugar donde guardo y analizo toda la información de tráfico y rendimiento de mi infraestructura Docker y Traefik.Te proporciono el código compose.yml para que puedas desplegar esta base de datos en cuestión de minutos y empezar a interactuar con los datos geolocalizados que envía Vector.Además de la solución Vector/OpenObserve, te presento un interesante descubrimiento: el Traefik Log Dashboard. Este proyecto de código abierto (backend en Go, frontend en React) te permite tener información en tiempo real de los logs de Traefik con geolocalización incluida.Monitorización en tiempo real vía WebSocket.Soporte para trazas en tiempo real (OpenTelemetry OTLP).Analíticas completas de tiempos de respuesta, códigos de estado y tasas de solicitud.Más información y enlaces en las notas del episodio
A revista TIME fez uma lista, o Rambo foi ali e já voltou, e o Coca manja muito de Docker.
We join a fair number of projects, and we often help teams bring their project up to our standard. This means bringing a lot of the same small pieces from project to project.In the latest episode of the No Compromises podcast, we rethink our “project standard” repo. Instead of a full Laravel skeleton, we propose a composable library of tool-specific, versioned configs (PHPUnit, Docker, etc.). We walk through the benefits for greenfield and legacy work, open questions about test organization, and how this approach scales as tools evolve.(00:00) - Why we keep our tooling current (00:15) - The “project standard” repo is aging (01:30) - Reference guide vs installable skeleton (02:30) - Supporting old and new stacks (versions, tags) (03:30) - Pivot: organize by tool and version, not app (04:30) - Example plan: folders for PHPUnit 11/12 (and beyond) (05:15) - What belongs where? Tests, traits, and context (10:00) - Docker-first thinking; where Horizon config lives (11:15) - Open questions: PHPUnit vs Pest vs “testing” folder (12:15) - Takeaway: evolve the repo as the tools evolve (12:45) - Silly bit Want help making your project as organized as one of our projects?
Bret is joined by Philip Andrews and Dan Muret of Cast AI to discuss pod live migration between nodes in a Kubernetes cluster.
This week it's all about your feedback! We answer your questions, and dig through your problems! -- During The Show -- 00:50 Intro Hallucinating 03:45 Friso from Mumble Moving all chats to one application Matrix Server Bridges Beeper Messages not flowing Self hosting beeper bridges E2EE 12:05 Security Cameras - Dustin Used Axis cameras Reolink Camera's POE Ethernet only models Some models require Reolink app for first setup Doorbell camera Steve's "doorbell camera flow" Surveillance Station Get one drive Reduce frame rate Weeks worth of footage Axis camera 2035 or 2036 Axis P3367 Write back in! 24:04 Laptops - Dasgeek Framework 12 Framework 12 vs Framework 13 25:45 News Wire Proxmox Mail Gateway 9.0 - helpnetsecurity.com (https://www.helpnetsecurity.com/2025/10/06/proxmox-mail-gateway-open-source-email-security-solution) GnuCash 5.13 - gnucash.org (https://www.gnucash.org/download.phtml) GNU Octave 10.3 - octave.org (https://octave.org/news/release/2025/10/01/octave-10.3.0-released.html) OpenSSL 3.6 - openssl-library.org (https://openssl-library.org/post/2025-10-01-3.6-release-announcement) Docker 28.5 - docker.com (https://docs.docker.com/engine/release-notes/28) Cairo-Dock 3.6 - phoronix.com (https://www.phoronix.com/news/Cairo-Dock-3.6-Released) GNU Linux Libre Kernel 6.17 - gnu.org (https://lists.gnu.org/archive/html/info-gnu/2025-09/msg00007.html) Tinycore 16.2 - tinycorelinux.net (https://forum.tinycorelinux.net/index.php/topic,27807.0.html) Opensuse Leap 16 - opensuse.org (https://get.opensuse.org/leap/16.0) Spine Endoscopic Atlas - nature.com (https://www.nature.com/articles/s41597-025-05897-7) Huawei Shrinks LLMs - venturebeat.com (https://venturebeat.com/ai/huaweis-new-open-source-technique-shrinks-llms-to-make-them-run-on-less) 4th Version of Granite AI - techrepublic.com (https://www.techrepublic.com/article/news-ibm-granite-40-ai) Granite AI Earns ISO 42001 - ibm.com (https://www.ibm.com/new/announcements/ibm-granite-iso-42001) 27:00 Send Feedback and Questions! Newbie questions welcome Fastest way to learn is immerse yourself Experts welcome too Email in! Follow up 30:45 Linux Basics - Kevin The bell story Level of learning GUI vs CLI Learning styles Ask your Linux friends Ask a LLM SIMPLE things Get hands on using it Set your expectations Tails Linux (https://tails.net/) Reproducibility Endless OS (https://www.endlessglobal.com/) Distro Hop Fedora (https://fedoraproject.org//) Ubuntu Arch Linux (https://archlinux.org/) Tiling window manager RHCSA and other Red Hat exams Canonical Academy ANS 437 (https://podcast.asknoahshow.com/437) -- The Extra Credit Section -- For links to the articles and material referenced in this week's episode check out this week's page from our podcast dashboard! This Episode's Podcast Dashboard (http://podcast.asknoahshow.com/462) Phone Systems for Ask Noah provided by Voxtelesys (http://www.voxtelesys.com/asknoah) Join us in our dedicated chatroom #GeekLab:linuxdelta.com on Matrix (https://element.linuxdelta.com/#/room/#geeklab:linuxdelta.com) -- Stay In Touch -- Find all the resources for this show on the Ask Noah Dashboard Ask Noah Dashboard (http://www.asknoahshow.com) Need more help than a radio show can offer? Altispeed provides commercial IT services and they're excited to offer you a great deal for listening to the Ask Noah Show. Call today and ask about the discount for listeners of the Ask Noah Show! Altispeed Technologies (http://www.altispeed.com/) Contact Noah live [at] asknoahshow.com -- Twitter -- Noah - Kernellinux (https://twitter.com/kernellinux) Ask Noah Show (https://twitter.com/asknoahshow) Altispeed Technologies (https://twitter.com/altispeed)
Shawn Tierney meets up with Connor Mason of Software Toolbox to learn their company, products, as well as see a demo of their products in action in this episode of The Automation Podcast. For any links related to this episode, check out the “Show Notes” located below the video. Watch The Automation Podcast from The Automation Blog: Listen to The Automation Podcast from The Automation Blog: The Automation Podcast, Episode 248 Show Notes: Special thanks to Software Toolbox for sponsoring this episode so we could release it “ad free!” To learn about Software Toolbox please checkout the below links: TOP Server Cogent DataHub Industries Case studies Technical blogs Read the transcript on The Automation Blog: (automatically generated) Shawn Tierney (Host): Welcome back to the automation podcast. My name is Shawn Tierney with Insights and Automation, and I wanna thank you for tuning back in this week. Now this week on the show, I meet up with Connor Mason from Software Toolbox, who gives us an overview of their product suite, and then he gives us a demo at the end. And even if you’re listening, I think you’re gonna find the demo interesting because Connor does a great job of talking through what he’s doing on the screen. With that said, let’s go ahead and jump into this week’s episode with Connor Mason from Software Toolbox. I wanna welcome Connor from Software Toolbox to the show. Connor, it’s really exciting to have you. It’s just a lot of fun talking to your team as we prepared for this, and, I’m really looking forward to because I just know in your company over the years, you guys have so many great solutions that I really just wanna thank you for coming on the show. And before you jump into talking about products and technologies Yeah. Could you first tell us just a little bit about yourself? Connor Mason (Guest): Absolutely. Thanks, Shawn, for having us on. Definitely a pleasure to be a part of this environment. So my name is Connor Mason. Again, I’m with Software Toolbox. We’ve been around for quite a while. So we’ll get into some of that history as well before we get into all the the fun technical things. But, you know, I’ve worked a lot with the variety of OT and IT projects that are ongoing at this point. I’ve come up through our support side. It’s definitely where we grow a lot of our technical skills. It’s a big portion of our company. We’ll get that into that a little more. Currently a technical application consultant lead. So like I said, I I help run our support team, help with these large solutions based projects and consultations, to find what’s what’s best for you guys out there. There’s a lot of different things that in our in our industry is new, exciting. It’s fast paced. Definitely keeps me busy. My background was actually in data analytics. I did not come through engineering, did not come through the automation, trainings at all. So this is a whole new world for me about five years ago, and I’ve learned a lot, and I really enjoyed it. So, I really appreciate your time having us on here, Shawn Tierney (Host): Shawn. Well, I appreciate you coming on. I’m looking forward to what you’re gonna show us today. I had a the audience should know I had a little preview of what they were gonna show, so I’m looking forward to it. Connor Mason (Guest): Awesome. Well, let’s jump right into it then. So like I said, we’re here at Software Toolbox, kinda have this ongoing logo and and just word map of connect everything, and that’s really where we lie. Some people have called us data plumbers in the past. It’s all these different connections where you have something, maybe legacy or something new, you need to get into another system. Well, how do you connect all those different points to it? And, you know, throughout all these projects we worked on, there’s always something unique in those different projects. And we try to work in between those unique areas and in between all these different integrations and be something that people can come to as an expert, have those high level discussions, find something that works for them at a cost effective solution. So outside of just, you know, products that we offer, we also have a lot of just knowledge in the industry, and we wanna share that. You’ll kinda see along here, there are some product names as well that you might recognize. Our top server and OmniServer, we’ll be talking about LOPA as well. It’s been around in the industry for, you know, decades at this point. And also our symbol factory might be something you you may have heard in other products, that they actually utilize themselves for HMI and and SCADA graphics. That is that is our product. So you may have interacted it with us without even knowing it, and I hope we get to kind of talk more about things that we do. So before we jump into all the fun technical things as well, I kind of want to talk about just the overall software toolbox experience as we call it. We’re we’re more than just someone that wants to sell you a product. We we really do work with, the idea of solutions. How do we provide you value and solve the problems that you are facing as the person that’s actually working out there on the field, on those operation lines, and making things as well. And that’s really our big priority is providing a high level of knowledge, variety of the things we can work with, and then also the support. It’s very dear to me coming through the the support team is still working, you know, day to day throughout that software toolbox, and it’s something that has been ingrained into our heritage. Next year will be thirty years of software toolbox in 2026. So we’re established in 1996. Through those thirty years, we have committed to supporting the people that we work with. And I I I can just tell you that that entire motto lives throughout everyone that’s here. So from that, over 97% of the customers that we interact with through support say they had an awesome or great experience. Having someone that you can call that understands the products you’re working with, understands the environment you’re working in, understands the priority of certain things. If you ever have a plant shut down, we know how stressful that is. Those are things that we work through and help people throughout. So this really is the core pillars of Software Toolbox and who we are, beyond just the products, and and I really think this is something unique that we have continued to grow and stand upon for those thirty years. So jumping right into some of the industry challenges we’ve been seeing over the past few years. This is also a fun one for me, talking about data analytics and tying these things together. In my prior life and education, I worked with just tons of data, and I never fully knew where it might have come from, why it was such a mess, who structured it that way, but it’s my job to get some insights out of that. And knowing what the data actually was and why it matters is a big part of actually getting value. So if you have dirty data, if you have data that’s just clustered, it’s in silos, it’s very often you’re not gonna get much value out of it. This was a study that we found in 2024, from Garner Research, And it said that, based on the question that business were asked, were there any top strategic priorities for your data analytics functions in 2024? And almost 50%, it’s right at ’49, said that they wanted to improve data quality, and that was a strategic priority. This is about half the industry is just talking about data quality, and it’s exactly because of those reasons I said in my prior life gave me a headache, to look at all these different things that I don’t even know where they became from or or why they were so different. And the person that made that may have been gone may not have the contacts, and making that from the person that implemented things to the people that are making decisions, is a very big task sometimes. So if we can create a better pipeline of data quality at the beginning, makes those people’s lives a lot easier up front and allows them to get value out of that data a lot quicker. And that’s what businesses need. Shawn Tierney (Host): You know, I wanna just data quality. Right? Mhmm. I think a lot of us, when we think of that, we think of, you know, error error detection. We think of lost connections. We think of, you know, just garbage data coming through. But I I think from an analytical side, there’s a different view on that, you know, in line with what you were just saying. So how do you when you’re talking to somebody about data quality, how do you get them to shift gears and focus in on what you’re talking about and not like a quality connection to the device itself? Connor Mason (Guest): Absolutely. Yeah. We I kinda live in both those worlds now. You know, I I get to see that that connection state. And when you’re operating in real time, that quality is also very important to you. Mhmm. And I kind of use that at the same realm. Think of that when you’re thinking in real time, if you know what’s going on in the operation and where things are running, that’s important to you. That’s the quality that you’re looking for. You have to think beyond just real time. We’re talking about historical data. We’re talking about data that’s been stored for months and years. Think about the quality of that data once it’s made up to that level. Are they gonna understand what was happening around those periods? Are they gonna understand what those tags even are? Are they gonna understand what those conventions that you’ve implemented, to give them insights into this operation. Is that a clear picture? So, yeah, you’re absolutely right. There are two levels to this, and and that is a big part of it. The the real time data and historical, and we’re gonna get some of that into into our demo as well. It it’s a it’s a big area for the business, and the people working in the operations. Shawn Tierney (Host): Yeah. I think quality too. Think, you know, you may have data. It’s good data. It was collected correctly. You had a good connection to the device. You got it. You got it as often as you want. But that data could really be useless. It could tell you nothing. Connor Mason (Guest): Right. Exactly. Shawn Tierney (Host): Right? It could be a flow rate on part of the process that irrelevant to monitoring the actual production of the product or or whatever you’re making. And, you know, I’ve known a lot of people who filled up their databases, their historians, with they just they just logged everything. And it’s like a lot of that data was what I would call low quality because it’s low information value. Right? Absolutely. I’m sure you run into that too. Connor Mason (Guest): Yeah. We we run into a lot of people that, you know, I’ve got x amount of data points in my historian and, you know, then we start digging into, well, I wanna do something with it or wanna migrate. Okay. Like, well, what do you wanna achieve at the end of this? Right? And and asking those questions, you know, it’s great that you have all these things historized. Are you using it? Do you have the right things historized? Are they even set up to be, you know, worked upon once they are historized by someone outside of this this landscape? And I think OT plays such a big role in this, and that’s why we start to see the convergence of the IT and OT teams just because that communication needs to occur sooner. So we’re not just passing along, you know, low quality data, bad quality data as well. And we’ll get into some of that later on. So to jump into some of our products and solutions, I kinda wanna give this overview of the automation pyramid. This is where we work from things like the field device communications. And you you have certain sensors, meters, actuators along the actual lines, wherever you’re working. We work across all the industries, so this can vary between those. Through there, you work up kind of your control area. A lot of control engineers are working. This is where I think a lot of the audience is very familiar with PLCs. Your your typical name, Siemens, Rockwell, your Schneiders that are creating, these hardware products. They’re interacting with things on the operation level, and they’re generating data. That that was kind of our bread and butter for a very long time and still is that communication level of getting data from there, but now getting it up the stack further into the pyramid of your supervisory, MES connections, and it’ll also now open to these ERP. We have a lot of large corporations that have data across variety of different solutions and also want to integrate directly down into their operation levels. There’s a lot of value to doing that, but there’s also a lot of watch outs, and a lot of security concerns. So that’ll be a topic that we’ll be getting into. We also all know that the cloud is here. It’s been here, and it’s it’s gonna continue to push its way into, these cloud providers into OT as well. There there’s a lot of benefit to it, but there there’s also some watch outs as this kind of realm, changes in the landscape that we’ve been used to. So there’s a lot of times that we wanna get data out there. There’s value into AI agents. It’s a hot it’s a hot commodity right now. Analytics as well. How do we get those things directly from shop floor, up into the cloud directly, and how do we do that securely? It’s things that we’ve been working on. We’ve had successful projects, continues to be an interest area and I don’t see it slowing down at all. Now, when we kind of begin this level at the bottom of connectivity, people mostly know us for our top server. This is our platform for industrial device connectivity. It’s a thing that’s talking to all those different PLCs in your plant, whether that’s brownfield or greenfield. We pretty much know that there’s never gonna be a plant that’s a single PLC manufacturer, that exists in one plant. There’s always gonna be something that’s slightly different. Definitely from Brownfield, things different engineers made different choices, things have been eminent, and you gotta keep running them. TopServe provides this single platform to connect to a long laundry list of different PLCs. And if this sounds very familiar to Kepserver, well, you’re not wrong. Kepserver is the same exact technology that TopServer is. What’s the difference then is probably the biggest question we usually get. The difference technology wise is nothing. The difference in the back end is that actually it’s all the same product, same product releases, same price, but we have been the biggest single source of Kepserver or Topsyra implementation into the market, for almost two plus decades at this point. So the single biggest purchase that we own this own labeled version of Kepserver to provide to our customers. They interact with our support team, our solutions teams as well, and we sell it along the stack of other things because it it fits so well. And we’ve been doing this since the early two thousands when, Kepware was a a much smaller company than it is now, and we’ve had a really great relationship with them. So if you’ve enjoyed the technology of of Kepserver, maybe there’s some users out there. If you ever heard of TopServer and that has been unclear, I hope this clear clarifies it. But it it is a great technology stack that that we build upon and we’ll get into some of that in our demo. Now the other question is, what if you don’t have a standard communication protocol, like a modbus, like an Allen Bradley PLC as well? We see this a lot with, you know, testing areas, pharmaceuticals, maybe also in packaging, barcode scanners, weigh scales, printers online as well. They they may have some form of basic communications that talks over just TCP or or serial. And how do you get that information that’s really valuable still, but it’s not going through a PLC. It’s not going into your typical agent mind SCADA. It might be very manual process for a lot of these test systems as well, how they’re collecting and analyzing the data. Well, you may have heard of our Arm server as well. It’s been around, like I said, for a couple decades and just a proven solution that without coding, you can go in and build a custom protocol that expects a format from that device, translates it, puts it into standard tags, and now that those tags can be accessible through the open standards of OPC, or to it was a a Veeva user suite link as well. And that really provides a nice combination of your standard communications and also these more custom communications may have been done through scripting in the past. Well, you know, put this onto, an actual server that can communicate through those protocols natively, and just get that data into those SCADA systems, HMIs, where you need it. Shawn Tierney (Host): You know, I used that. Many years ago, I had an integrator who came to me. He’s like, Shawn, I wanna this is back in the RSVUE days. He’s like, Shawn, I I got, like, 20 Euotherm devices on a four eighty five, and they speak ASCII, and I gotta I gotta get into RSVUE 32. And, you know, OmniSIR, I love that you could you could basically developing and we did Omega and some other devices too. You’re developing your own protocol, but it’s beautiful. And and the fact that when you’re testing it, it color codes everything. So you know, hey. That part worked. The header worked. The data worked. Oh, the trailing didn’t work, or the terminated didn’t work, or the data’s not in the right format. Or I just it was a joy to work with back then, and I can imagine it’s only gotten better since. Connor Mason (Guest): Yeah. I think it’s like a little engineer playground where you get in there. It started really decoding and seeing how these devices communicate. And then once you’ve got it running, it it’s one of those things that it it just performs and, is saved by many people from developing custom code, having to manage that custom code and integrations, you know, for for many years. So it it’s one of those things that’s kinda tried, tested, and, it it’s kind of a staple still our our base level communications. Alright. So moving along kind of our automation pyramid as well. Another part of our large offering is the Cogent data hub. Some people may have heard from this as well. It’s been around for a good while. It’s been part of our portfolio for for a while as well. This starts building upon where we had the communication now up to those higher echelons of the pyramid. This is gonna bring in a lot of different connectivities. You if you’re not if you’re listening, it it’s kind of this cog and spoke type of concept for real time data. We also have historical implementations. You can connect through a variety of different things. OPC, both the profiles for alarms and events, and even OPC UA’s alarming conditions, which is still getting adoption across the, across the industry, but it is growing. As part of the OPC UA standard, we have integrations to MQTT. It can be its own MQTT broker, and it can also be an MQTT client. That has grown a lot. It’s one of those things that lives be besides OPC UA, not exactly a replacement. If you ever have any questions about that, it’s definitely a topic I love to talk about. There’s space for for this to combine the benefits of both of these, and it’s so versatile and flexible for these different type of implementations. On top of that, it it’s it’s a really strong tool for conversion and aggregation. You kind of add this, like, its name says, it’s a it’s a data hub. You send all the different information to this. It stores it into, a hierarchy with a variety of different modeling that you can do within it. That’s gonna store these values across a standard data format. Once I had data into this, any of those different connections, I can then send data back out. So if I have anything that I know is coming in through a certain plug in like OPC, bring that in, send it out to on these other ones, OPC, DA over to MQTT. It could even do DDA if I’m still using that, which I probably wouldn’t suggest. But overall, there’s a lot of good benefits from having something that can also be a standardization, between all your different connections. I have a lot of different things, maybe variety of OPC servers, legacy or newer. Bring that into a data hub, and then all your other connections, your historians, your MAS, your SCADAs, it can connect to that single point. So it’s all getting the same data model and values from a single source rather than going out and making many to many connections. A a large thing that it was originally, used for was getting around DCOM. That word is, you know, it might send some shivers down people’s spines still, to this day, but it’s it’s not a fun thing to deal with DCOM and also with the security hardening. It’s just not something that you really want to do. I’m sure there’s a lot of security professionals would advise against EPRA doing it. This tunneling will allow you to have a data hub that locally talks to any of the DA server client, communicate between two data hubs over a tunnel that pushes the data just over TCP, takes away all the comm wrappers, and now you just have values that get streamed in between. Now you don’t have to configure any DCOM at all, and it’s all local. So a lot of people went transitioning, between products where maybe the server only supports OPC DA, and then the client is now supporting OPC UA. They can’t change it yet. This has allowed them to implement a solution quickly and cost and at a cost effective price, without ripping everything out. Shawn Tierney (Host): You know, I wanna ask you too. I can see because this thing is it’s a data hub. So if you’re watching and you’re if you’re listening and not watching, you you’re not gonna see, you know, server, client, UAD, a broker, server, client. You know, just all these different things up here on the site. Do you what how does somebody find out if it does what they need? I mean, do you guys have a line they can call to say, I wanna do this to this. Is that something Data Hub can do, or is there a demo? What would you recommend to somebody? Connor Mason (Guest): Absolutely. Reach out to us. We we have a a lot of content outline, and it’s not behind any paywall or sign in links even. You you can always go to our website. It’s just softwaretoolbox.com. Mhmm. And that’s gonna get you to our product pages. You can download any product directly from there. They have demo timers. So typically with, with coaching data hub, after an hour, it will stop. You can just rerun it. And then call our team. Yeah. We have a solutions team that can work with you on, hey. What do I need as well? Then our support team, if you run into any issues, can help you troubleshoot that as well. So, I’ll have some contact information at the end, that’ll get some people to, you know, where they need to go. But you’re absolutely right, Shawn. Because this is so versatile, everyone’s use case of it is usually something a little bit different. And the best people to come talk to that is us because we’ve we’ve seen all those differences. So Shawn Tierney (Host): I think a lot of people run into the fact, like, they have a problem. Maybe it’s the one you said where they have the OPC UA and it needs to connect to an OPC DA client. And, you know, and a lot of times, they’re they’re a little gunshot to buy a license because they wanna make sure it’s gonna do exactly what they need first. And I think that’s where having your people can, you know, answer their questions saying, yes. We can do that or, no. We can’t do that. Or, you know, a a demo that they could download and run for an hour at a time to actually do a proof of concept for the boss who’s gonna sign off on purchasing this. And then the other thing is too, a lot of products like this have options. And you wanna make sure you’re buying the ticking the right boxes when you buy your license because you don’t wanna buy something you’re not gonna use. You wanna buy the exact pieces you need. So I highly recommend I mean, this product just does like, I have, in my mind, like, five things I wanna ask right now, but not gonna. But, yeah, def definitely, when it when it comes to a product like this, great to touch base with these folks. They’re super friendly and helpful, and, they’ll they’ll put you in the right direction. Connor Mason (Guest): Yeah. I I can tell you that’s working someone to support. Selling someone a solution that doesn’t work is not something I’ve been doing. Bad day. Right. Exactly. Yeah. And we work very closely, between anyone that’s looking at products. You know, me being as technical product managers, well, I I’m engaged in those conversations. And Mhmm. Yeah. If you need a demo license, reach out to us to extend that. We wanna make sure that you are buying something that provides you value. Now kind of moving on into a similar realm. This is one of our still somewhat newer offerings, I say, but we’ve been around five five plus years, and it’s really grown. And I kinda said here, it’s called OPC router, and and it’s not it’s not a networking tool. A lot of people may may kinda get that. It’s more of a, kind of a term about, again, all these different type of connections. How do you route them to different ways? It it kind of it it separates itself from the Cogent data hub, and and acting at this base level of being like a visual workflow that you can assign various tasks to. So if I have certain events that occur, I may wanna do some processing on that before I just send data along, where the data hub is really working in between converting, streaming data, real time connections. This gives you a a kind of a playground to work around of if I have certain tasks that are occurring, maybe through a database that I wanna trigger off of a certain value, based on my SCADA system, well, you can build that in in these different workflows to execute exactly what you need. Very, very flexible. Again, it has all these different type of connections. The very unique ones that have also grown into kind of that OT IT convergence, is it can be a REST API server and client as well. So I can be sending out requests to, RESTful servers where we’re seeing that hosted in a lot of new applications. I wanna get data out of them. Or once I have consumed a variety of data, I can become the REST server in OPC router and offer that to other applications to request data from itself. So, again, it can kind of be that centralized area of information. The other thing as we talked about in the automation pyramid is it has connections directly into SAP and ERP systems. So if you have work orders, if you have materials, that you wanna continue to track and maybe trigger things based off information from your your operation floors via PLCs tracking, how they’re using things along the line, and that needs to match up with what the SAP system has for, the amount of materials you have. This can be that bridge. It’s really is built off the mindset of the OT world as well. So we kinda say this helps empower the OT level because we’re now giving them the tools to that they understand what what’s occurring in their operations. And what could you do by having a tool like this to allow you to kind of create automated workflows based off certain values and certain events and automate some of these things that you may be doing manually or doing very convoluted through a variety of solutions. So this is one of those prod, products as well that’s very advanced in the things that supports. Linux and Docker containers is, is definitely could be a hot topic, rightly fleet rightfully so. And this can run on a on a Docker container deployed as well. So we we’ve seen that with the I IT folks that really enjoy being able to control and to higher deployment, allows you to update easily, allows you to control and spin up new containers as well. This gives you a lot of flexibility to to deploy and manage these systems. Shawn Tierney (Host): You know, I may wanna have you back on to talk about this. I used to there’s an old product called Rascal that I used to use. It was a transaction manager, and it would based on data changing or on a time that as a trigger, it could take data either from the PLC to the database or from the database to the PLC, and it would work with stored procedures. And and this seems like it hits all those points, And it sounds like it’s a visual like you said, right there on the slide, visual workflow builder. Connor Mason (Guest): Yep. Shawn Tierney (Host): So you really piqued my interest with this one, and and it may be something we wanna come back to and and revisit in the future, because, it just it’s just I know that that older product was very useful and, you know, it really solved a lot of old applications back in the day. Connor Mason (Guest): Yeah. Absolutely. And this this just takes that on and builds even more. If you if anyone was, kind of listening at the beginning of this year or two, a conference called Prove It that was very big in the industry, we were there to and we presented on stage a solution that we had. Highly recommend going searching for that. It’s on our web pages. It’s also on their YouTube links, and it’s it’s called Prove It. And OPC router was a big part of that in the back end. I would love to dive in and show you the really unique things. Kind of as a quick overview, we’re able to use Google AI vision to take camera data and detect if someone was wearing a hard hat. All that logic and behind of getting that information to Google AI vision, was through REST with OPC router. Then we were parsing that information back through that, connection and then providing it back to the PLCs. So we go all the way from a camera to a PLC controlling a light stack, up to Google AI vision through OPC router, all on hotel Wi Fi. It’s very imp it’s very, very fun presentation, and, our I think our team did a really great job. So a a a pretty new offering I have I wanna highlight, is our is our data caster. This is a an actual piece of hardware. You know, our software toolbox is we we do have some hardware as well. It’s just, part of the nature of this environment of how we mesh in between things. But the the idea is that, there’s a lot of different use cases for HMI and SCADA. They have grown so much from what they used to be, and they’re very core part of the automation stack. Now a lot of times, these are doing so many things beyond that as well. What we found is that in different areas of operations, you may not need all that different control. You may not even have the space to make up a whole workstation for that as well. What this does, the data caster, is, just simply plug it plugs it into any network and into an HDMI compatible display, and it gives you a very easy configure workplace to put a few key metrics onto a screen. So if I have different things from you can connect directly to PLCs like Allen Bradley. You can connect to SQL databases. You can also connect to rest APIs to gather the data from these different sources and build a a a kind of easy to to view, KPI dashboard in a way. So if you’re on a operation line and you wanna look at your current run rate, maybe you have certain things in the POC tags, you know, flow and pressure that’s very important for those operators to see. They may not be, even the capacity to be interacting with anything. They just need visualizations of what’s going on. This product can just be installed, you know, industrial areas with, with any type of display that you can easily access and and give them something that they can easily look at. It’s configured all through a web browser to display what you want. You can put on different colors based on levels of values as well. And it’s just I feel like a very simple thing that sometimes it seems so simple, but those might be the things that provide value on the actual operation floor. This is, for anyone that’s watching, kind of a quick view of a very simple screen. What we’re showing here is what it would look like from all the different data sources. So talking directly to ControlLogs PLC, talking to SQL databases, micro eight eight hundreds, an arrest client, and and what’s coming very soon, definitely by the end of this year, is OPC UA support. So any OPC UA server that’s out there that’s already having your PLC data or etcetera, this could also connect to that and get values from there. Shawn Tierney (Host): Can I can you make it I’m I’m here I go? Can you make it so it, like, changes, like, pages every few seconds? Connor Mason (Guest): Right now, it is a single page, but this is, like I said, very new product, so we’re taking any feedback. If, yeah, if there’s this type of slideshow cycle that would be, you know, valuable to anyone out there, let us know. We’re definitely always interested to see the people that are actually working out at these operation sites, what what’s valuable to them. Yeah. Shawn Tierney (Host): A lot of kiosks you see when when you’re traveling, it’ll say, like, line one well, I’ll just throw out there. Line one, and that’ll be on there for five seconds, and then it’ll go line two. That’ll be on there for five seconds, and then line you know, I and that’s why I just mentioned that because I can see that being a question that, that that I would get from somebody who is asking me about it. Connor Mason (Guest): Oh, great question. Appreciate it. Alright. So now we’re gonna set time for a little hands on demo. For anyone that’s just listening, we’re gonna I’m gonna talk about this at at a high level and walk through everything. But the idea is that, we have a few different POCs, very common in Allen Bradley and just a a Siemens seven, s seven fifteen hundred that’s in our office, pretty close to me on the other side of the wall wall, actually. We’re gonna first start by connecting that to our top server like we talked about. This is our industrial communication server, that offers both OCDA, OC UA, SweetLink connectivity as well. And then we’re gonna bring this into our Cogent data hub. This we talked about is getting those values up to these higher levels. What we’ll be doing is also tunneling the data. We talked about being able to share data through the data hubs themselves. Kinda explain why we’re doing that here and the value you can add. And then we’re also gonna showcase adding on MQTT to this level. Taking beta now just from these two PLCs that are sitting on a rack, and I can automatically make all that information available in the MQTT broker. So any MQTT client that’s out there that wants to subscribe to that data, now has that accessible. And I’ve created this all through a a really simple workflow. We also have some databases connected. Influx, we install with Code and DataHub, has a free visualization tool that kinda just helps you see what’s going on in your processes. I wanna showcase a little bit of that as well. Alright. So now jumping into our demo, when we first start off here is the our top server. Like I mentioned before, if anyone has worked with KEP server in the past, this is gonna look very similar. Like it because it is. The same technology and all the things here. The the first things that I wanted to establish in our demo, was our connection to our POCs. I have a few here. We’re only gonna use the Allen Bradley and the Siemens, for the the time that we have on our demo here. But how this builds out as a platform is you create these different channels and the devices connections between them. This is gonna be your your physical connections to them. It’s either, IP TCPIP connection or maybe your serial connection as well. We have support for all of them. It really is a long list. Anyone watching out there, you can kind of see all the different drivers that that we offer. So allowing this into a single platform, you can have all your connectivity based here. All those different connections that you now have that up the stack, your SCADA, your historians, MAS even as well, they can all go to a single source. Makes that management, troubleshooting, all those a bit easier as well. So one of the first things I did here, I have this built out, but I’ll kinda walk through what you would typically do. You have your Allen Bradley ControlLogix Ethernet driver here first. You know, I have some IPs in here I won’t show, but, regardless, we have our our our drivers here, and then we have a set of tags. These are all the global tags in the programming of the PLC. How I got these to to kind of map automatically is in our in our driver, we’re able to create tags automatically. So you’re able to send a command to that device and ask for its entire tag database. They can come back, provide all that, map it out for you, create those tags as well. This saves a lot of time from, you know, an engineer have to go in and, addressing all the individual items themselves. So once it’s defined in the program project, you’re able to bring this all in automatically. I’ll show now how easy that makes it connecting to something like the Cogent data hub. In a very similar fashion, we have a connection over here to the Siemens, PLC that I also have. You can see beneath it all these different tag structures, and this was created the exact same way. While those those PLC support it, you can do an automatic tag generation, bring in all the structure that you’ve already built out your PLC programming, and and make this available on this OPC server now as well. So that’s really the basis. We first need to establish communications to these PLCs, get that tag data, and now what do we wanna do with it? So in this demo, what I wanted to bring up was, the code in DataHub next. So here, I see a very similar kind of layout. We have a different set set of plugins on the left side. So for anyone listening, the Cogent Data Hub again is kind of our aggregation and conversion tool. All these different type of protocols like OPC UA, OPC DA, and OPC A and E for alarms and events. We also support OPC alarms and conditions, which is the newer profile for alarms in OPC UA. We have all a variety of different ways that you can get data out of things and data’s into the data hub. We can also do bridging. This concept is, how you share data in between different points. So let’s say I had a connection to one OPC server, and it was communicating to a certain PLC, and there were certain registers I was getting data from. Well, now I also wanna connect to a different OPC server that has, entirely different brand of PLCs. And then maybe I wanna share data in between them directly. Well, with this software, I can just bridge those points between them. Once they’re in the data hub, I can do kind of whatever I want with them. I can then allow them to write between those PLCs and share data that way, and you’re not now having to do any type of hardwiring directly in between them, and then I’m compatible to communicate to each other. Through the standards of OPC and these variety of different communication levels, I can integrate them together. Shawn Tierney (Host): You know, you bring up a good point. When you do something like that, is there any heartbeat? Like, is there on the general or under under, one of these, topics? Is there are there tags we can use that are from DataHub itself that can be sent to the destination, like a heartbeat or, you know, the merge transactions? Or Connor Mason (Guest): Yeah. Absolutely. So with this as well, there’s pretty strong scripting engine, and I have done that in the past where you can make internal tags. And that that could be a a timer. It could be a counter. And and just kind of allows you to create your own tags as well that you could do the same thing, could share that, through bridge connection to a PLC. So, yeah, there there are definitely some people that had those cert and, you know, use cases where they wanna get something to just track, on this software side and get it out to those hardware PLCs. Absolutely. Shawn Tierney (Host): I mean, when you send out the data out of the PLC, the PLC doesn’t care to take my data. But when you’re getting data into the PLC, you wanna make sure it’s updating and it’s fresh. And so, you know, they throw a counter in there, the script thing, and be able to have that. As as long as you see that incrementing, you know, you got good data coming in. That’s that’s a good feature. Connor Mason (Guest): Absolutely. You know, another big one is the the redundancy. So what this does is beyond just the OPC, we can make redundancy to basically anything that has two things running of it. So any of these different connections. How it’s unique is what it does is it just looks at the buckets of data that you create. So for an example, if I do have two different OPC servers and I put them into two areas of, let’s say, OPC server one and OPC server two, I can what now create an OPC redundancy data bucket. And now any client that connects externally to that and wants that data, it’s gonna go talk to that bucket of data. And that bucket of data is going to automatically change in between sources as things go down, things come back up, and the client would never know what’s hap what that happened unless you wanted to. There are internal tasks to show what’s the current source and things, but the idea is to make this trans kind of hidden that regardless of what’s going on in the operations, if I have this set up, I can have my external applications just reading from a single source without knowing that there’s two things behind it that are actually controlling that. Very important for, you know, historian connections where you wanna have a full complete picture of that data that’s coming in. If you’re able to make a redundant connection to two different, servers and then allow that historian to talk to a single point where it doesn’t have to control that switching back and forth. It it will just see that data flow streamlessly as as either one is up at that time. Kinda beyond that as well, there’s quite a few other different things in here. I don’t think we have time to cover all of them. But for for our demo, what I wanna focus on first is our OPC UA connection. This allows us both to act as a OPC UA client to get data from any servers out there, like our top server. And also we can act as an OPC UA server itself. So if anything’s coming in from maybe you have multiple connections to different servers, multiple connections to other things that aren’t OPC as well, I can now provide all this data automatically in my own namespace to allow things to connect to me as well. And that’s part of that aggregation feature, and kind of topic I was mentioning before. So with that, I have a connection here. It’s pulling data all from my top server. I have a few different tags from my Alec Bradley and and my Siemens PLC selected. The next part of this, while I was meshing, was the tunneling. Like I said, this is very popular to get around DCOM issues, but there’s a lot of reasons why you still may use this beyond just the headache of DCOM and what it was. What this runs on is a a TCP stream that takes all the data points as a value, a quality, and a timestamp, and it can mirror those in between another DataHub instance. So if I wanna get things across a network, like my OT side, where NASH previously, I would have to come in and allow a, open port onto my network for any OPC UA clients, across the network to access that, I can now actually change the direction of this and allow me to tunnel data out of my network without opening up any ports. This is really big for security. If anyone out there, security professional or working as an engineer, you have to work with your IT and security a lot, they don’t you don’t wanna have an open port, especially to your operations and OT side. So this allows you to change that direction of flow and push data out of this direction into another area like a DMZ computer or up to a business level computer as well. The other things as well that I have configured in this demo, the benefit of having that tunneling streaming data across this connection is I can also store this data locally in a, influx database. The purpose of that then is that I can actually historize this, provide then if this connection ever goes down to backfill any information that was lost during that tunnel connection going down. So with this added layer on and real time data scenarios like OPC UA, unless you have historical access, you would lose a lot of data if that connection ever went down. But with this, I can actually use the back end of this InfluxDB, buffer any values. When my connection comes back up, pass them along that stream again. And if I have anything that’s historically connected, like, another InfluxDB, maybe a PI historian, Vue historian, any historian offering out there that can allow that connection. I can then provide all those records that were originally missed and backfill that into those systems. So I switched over to a second machine. It’s gonna look very similar here as well. This also has an instance of the Cogent Data Hub running here. For anyone not watching, what we’ve actually have on this side is the the portion of the tunneler that’s sitting here and listening for any data requests coming in. So on my first machine, I was able to connect my PLCs, gather that information into Cogent DataHub, and now I’m pushing that information, across the network into a separate machine that’s sitting here and listening to gather information. So what I can quickly do is just make sure I have all my data here. So I have these different points, both from my Allen Bradley PLCs. I have a few, different simulation demo points, like temperature, pressure, tank level, a few statuses, and all this is updating directly through that stream as the PLC is updating it as well. I also have my scenes controller. I have some, current values and a few different counters tags as well. All of this again is being directly streamed through that tunnel. I’m not connecting to an OPC server at all on this side. I can show you that here. There’s no connections configured. I’m not talking to the PLCs directly on this machine as well. But maybe we’ll pass all the information through without opening up any ports on my OT demo machine per se. So what’s the benefit of that? Well, again, security. Also, the ability to do the store and forward mechanisms. On the other side, I was logging directly to a InfluxDB. This could be my d- my buffer, and then I was able to configure it where if any values were lost, to store that across the network. So now with this side, if I pull up Chronic Graph, which is a free visualization tool that installs with the DataHub as well, I can see some very nice, visual workflows and and visual diagrams of what is going on with this data. So I have a pressure that is just a simulator in this, Allen Bradley PLC that ramps up and and comes back down. It’s not actually connected to anything that’s reading a real pressure, but you can see over time, I can kind of change through these different layers of time. And I might go back a little far, but I have a lot of data that’s been stored in here. For a while during my test, I turned this off and, made it fail, but then I came back in and I was able to recreate all the data and backfill it as well. So through through these views, I can see that as data disconnects, as it comes back on, I have a very cyclical view of the data because it was able to recover and store and forward from that source. Like I said, Shawn, data quality is a big thing in this industry. It’s a big thing for people both at the operations side, and both people making decision in the business layer. So being able to have a full picture, without gaps, it is definitely something that, you should be prioritizing, when you can. Shawn Tierney (Host): Now what we’re seeing here is you’re using InfluxDB on this, destination PC or IT side PC and chronograph, which was that utility or that package that comes, gets installed. It’s free. But you don’t actually have to use that. You could have sent this in to an OSI pi or Exactly. Somebody else’s historian. Right? Can you name some of the historians you work with? I know OSI pie. Connor Mason (Guest): Yeah. Yeah. Absolutely. So there’s quite a few different ones. As far as what we support in the Data Hub natively, Amazon Kinesis, the cloud hosted historian that we can also do the same things from here as well. Aviva Historian, Aviva Insight, Apache Kafka. This is a a kind of a a newer one as well that used to be a very IT oriented solution, now getting into OT. It’s kind of a similar database structure where things are stored in different topics that we can stream to. On top of that, just regular old ODBC connections. That opens up a lot of different ways you can do it, or even, the old classic OPC, HDA. So if you have any, historians that that can act as an OPC HDA, connection, we we can also stream it through there. Shawn Tierney (Host): Excellent. That’s a great list. Connor Mason (Guest): The other thing I wanna show while we still have some time here is that MQTT component. This is really growing and, it’s gonna continue to be a part of the industrial automation technology stack and conversations moving forward, for streaming data, you know, from devices, edge devices, up into different layers, both now into the OT, and then maybe out to, IT, in our business levels as well, and definitely into the cloud as we’re seeing a lot of growth into it. Like I mentioned with Data Hub, the big benefit is I have all these different connections. I can consume all this data. Well, I can also act as an MQTT broker. And what what a broker typically does in MQTT is just route data and share data. It’s kind of that central point where things come to it to either say, hey. I’m giving you some new values. Share it with someone else. Or, hey. I need these values. Can you give me that? It really fits in super well with what this product is at its core. So all I have to do here is just enable it. What that now allows is I have an example, MQTT Explorer. If anyone has worked with MQTT, you’re probably familiar with this. There’s nothing else I configured beyond just enabling the broker. And you can see within this structure, I have all the same data that was in my Data Hub already. The same things I were collecting from my PLCs and top server. Now I’ve embedded these as MPPT points and now I have them in JSON format with the value, their timestamp. You can even see, like, a little trend here kind of matching what we saw in Influx. And and now this enables all those different cloud connectors that wanna speak this language to do it seamlessly. Shawn Tierney (Host): So you didn’t have to set up the PLCs a second time to do this? Nope. Connor Mason (Guest): Not at all. Shawn Tierney (Host): You just enabled this, and now the data’s going this way as well. Exactly. Connor Mason (Guest): Yeah. That’s a really strong point of the Cogent Data Hub is once you have everything into its structure and model, you just enable it to use any of these different connections. You can get really, really creative with these different things. Like we talked about with the the bridging aspect and getting into different systems, even writing down the PLCs. You can make crust, custom notifications and email alerts, based on any of these values. You could even take something like this MTT connection, tunnel it across to another data hub as well, maybe then convert it to OPC DA. And now you’ve made a a a new connection over to something that’s very legacy as well. Shawn Tierney (Host): Yeah. That, I mean, the options here are just pretty amazing, all the different things that can be done. Connor Mason (Guest): Absolutely. Well, I, you know, I wanna jump back into some of our presentation here while we still got the time. And now after we’re kinda done with our demo, there’s so many different ways that you can use these different tools. This is just a really simple, kind of view of the, something that used to be very simple, just connecting OpenSea servers to a variety of different connections, kind of expanding onto with that that’s store and forward, the local influx usage, getting out to things like MTT as well. But there’s a lot more you can do with these solutions. So like Shawn said, reach out to us. We’re happy to engage and see what we can help you with. I have a few other things before we wrap up. Just overall, it we’ve worked across nearly every industry. We have installations across the globe on all continents. And like I said, we’ve been around for pushing thirty years next year. So we’ve seen a lot of different things, and we really wanna talk to anyone out there that maybe has some struggles that are going on with just connectivity, or you have any ongoing projects. If you work in these different industries or if there’s nothing marked here and you have anything going on that you need help with, we’re very happy to sit down and let you know if there’s there’s something we can do there. Shawn Tierney (Host): Yeah. For those who are, listening, I mean, we see most of the big energy and consumer product, companies on that slide. So I’m not gonna read them off, but, it’s just a lot of car manufacturers. You know, these are these are these, the household name brands that everybody knows and loves. Connor Mason (Guest): So kind of wrap some things up here. We talked about all the different ways that we’ve kind of helped solve things in the past, but I wanna highlight some of the unique ones, that we’ve also gone do some, case studies on and and success stories. So this one I actually got to work on, within the last few years that, a plastic packaging, manufacturer was looking to track uptime and downtime across multiple different lines, and they had a new cloud solution that they were already evaluating. They’re really excited to get into play. They they had a lot of upside to, getting things connected to this and start using it. Well, what they had was a lot of different PLCs, a lot of different brands, different areas, different, you know, areas of operation that they need to connect to. So what they used was to first get that into our top server, kind of similar to how they showed them use in their in our demo. We just need to get all the data into a centralized platform first, get that data accessible. Then from there, once they had all that information into a centralized area, they used the Cogent Data Hub as well to help aggregate that information and transform it to be sent to the cloud through MQTT. So very similar to the demo here, this is actually a real use case of that. Getting information from PLCs, structuring it into that how that cloud system needed it for MQTT, and streamlining that data connection to now where it’s just running in operation. They constantly have updates about where their lines are in operation, tracking their downtime, tracking their uptime as well, and then being able to do some predictive analytics in that cloud solution based on their history. So this really enabled them to kind of build from what they had existing. It was doing a lot of manual tracking, into an entirely automated system with management able to see real views of what’s going on at this operation level. Another one I wanna talk about was we we were able to do this success story with, Ace Automation. They worked with a pharmaceutical company. Ace Automation is a SI and they were brought in and doing a lot of work with some some old DDE connections, doing some custom Excel macros, and we’re just having a hard time maintaining some legacy systems that were just a pain to deal with. They were working with these older files, from some old InTouch histor HMIs, and what they needed to do was get something that was not just based on Excel and doing custom macros. So one product we didn’t get to talk about yet, but we also carry is our LGH file inspector. It’s able to take these files, put them out into a standardized format like CSV, and also do a lot of that automation of when when should these files be queried? Should they be, queried for different lengths? Should they be output to different areas? Can I set these up in a scheduled task so it can be done automatically rather than someone having to sit down and do it manually in Excel? So they will able to, recover over fifty hours of engineering time with the solution from having to do late night calls to troubleshoot a, Excel macro that stopped working, from crashing machines, because they were running a legacy systems to still support some of the DDE servers, into saving them, you know, almost two hundred plus hours of productivity. Another example, if we’re able to work with a renewable, energy customer that’s doing a lot of innovative things across North America, They had a very ambitious plan to double their footprint in the next two years. And with that, they had to really look back at their assets and see where they currently stand, how do we make new standards to support us growing into what we want to be. So with this, they had a lot of different data sources currently. They’re all kind of siloed at the specific areas. Nothing was really connected commonly to a corporate level area of historization, or control and security. So again, they they were able to use our top server and put out a standard connectivity platform, bring in the DataHub as an aggregation tool. So each of these sites would have a top server that was individually collecting data from different devices, and then that was able to send it into a single DataHub. So now their corporate level had an entire view of all the information from these different plants in one single application. That then enabled them to connect their historian applications to that data hub and have a perfect view and make visualizations off of their entire operations. What this allowed them to do was grow without replacing everything. And that’s a big thing that we try to strive on is replacing and ripping out all your existing technologies. It’s not something you can do overnight. But how do we provide value and gain efficiency with what’s in place and providing newer technologies on top of that without disrupting the actual operation as well? So this was really, really successful. And at the end, I just wanna kind of provide some other contacts and information people can learn more. We have a blog that goes out every week on Thursdays. A lot of good technical content out there. A lot of recast of the the awesome things we get to do here, the success stories as well, and you can always find that at justblog.softwaretoolbox.com. And again, our main website is justsoftwaretoolbox.com. You can get product information, downloads, reach out to anyone on our team. Let’s discuss what what issues you have going on, any new projects, we’ll be happy to listen. Shawn Tierney (Host): Well, Connor, I wanna thank you very much for coming on the show and bringing us up to speed on not only software toolbox, but also to, you know, bring us up to speed on top server and doing that demo with top server and data hub. Really appreciate that. And, I think, you know, like you just said, if anybody, has any projects that you think these solutions may be able to solve, please give them a give them a call. And if you’ve already done something with them, leave a comment. You know? To leave a comment, no matter where you’re watching or listening to this, let us know what you did. What did you use? Like me, I used OmniServer all those many years ago, and, of course, Top Server as an OPC server. But if you guys have already used Software Toolbox and, of course, Symbol Factory, I use that all the time. But if you guys are using it, let us know in the comments. It’s always great to hear from people out there. I know, you know, with thousands of you guys listening every week, but I’d love to hear, you know, are you using these products? Or if you have questions, I’ll funnel them over to Connor if you put them in the comments. So with that, Connor, did you have anything else you wanted to cover before we close out today’s show? Connor Mason (Guest): I think that was it, Shawn. Thanks again for having us on. It was really fun. Shawn Tierney (Host): I hope you enjoyed that episode, and I wanna thank Connor for taking time out of his busy schedule to come on the show and bring us up to speed on software toolbox and their suite of products. Really appreciated that demo at the end too, so we actually got a look at if you’re watching. Gotta look at their products and how they work. And, just really appreciate them taking all of my questions. I also appreciate the fact that Software Toolbox sponsored this episode, meaning we were able to release it to you without any ads. So I really appreciate them. If you’re doing any business with Software Toolbox, please thank them for sponsoring this episode. And with that, I just wanna wish you all good health and happiness. And until next time, my friends, peace. Until next time, Peace ✌️ If you enjoyed this content, please give it a Like, and consider Sharing a link to it as that is the best way for us to grow our audience, which in turn allows us to produce more content
In this episode of the Ardan Labs Podcast, Bill Kennedy talks with Mike Elgan, technology editorialist, about his journey to becoming a professional writer. Mike reflects on the evolution of journalism, the digital publishing revolution of the 90s, and the role of communication in shaping society. He also shares insights from his nomadic lifestyle, the realities of freelance writing, the impact of AI on creativity, and the importance of adaptability and passion in navigating a rapidly changing technological landscape.00:00 Introduction 05:54 Growing Up in the 70s and 80s11:32 Education, Passion & Early Work31:32 Politics, Journalism & First Writing Roles45:16 Transition to Technology Journalism54:30 Digital Media Revolution01:03:01 Becoming a Freelance Writer01:05:52 Nomadic Lifestyle & Financial Realities01:24:36 AI's Impact on Work, Writing & Education01:42:56 The Future of Writing & Communication01:50:52 Cyberpunk Reality & Technology's Next Era01:55:11 Finding Passion & Success in LifeConnect with Mike: Linkedin: https://www.linkedin.com/in/mikeelganX: https://x.com/MikeElganMentioned in this Episode:Mike Elgan's Newsletter: https://machinesociety.ai/Gastronomad: https://gastronomad.netWant more from Ardan Labs? You can learn Go, Kubernetes, Docker & more through our video training, live events, or through our blog!Online Courses : https://ardanlabs.com/education/ Live Events : https://www.ardanlabs.com/live-training-events/ Blog : https://www.ardanlabs.com/blog Github : https://github.com/ardanlabs
Alarm bells are ringing over a supposed browser zero-day, but is the threat as bad as it sounds? Steve reveals why "clickjacking" might be more whac-a-mole than breaking news, and what that really means for your passwords. • Germany may soon outlaw ad blockers • What's happening in the courts over AI • The U.K. drops its demands of Apple • New Microsoft 365 tenants being throttled • Is Russia preparing to block Google Meet? • Bluesky suspends its service in Mississippi • How to throttle AI • A tricky SSH-busting Go library • Here comes the Linux desktop malware • Apple just patched a doozy of a vulnerability • A trivial Docker escape was found and fixed • Why the recent browser 0-day clickjacking is really just whac-a-mole Show Notes - https://www.grc.com/sn/sn-1040-notes.pdf Hosts: Steve Gibson and Leo Laporte Download or subscribe to Security Now at https://twit.tv/shows/security-now. You can submit a question to Security Now at the GRC Feedback Page. For 16kbps versions, transcripts, and notes (including fixes), visit Steve's site: grc.com, also the home of the best disk maintenance and recovery utility ever written Spinrite 6. Join Club TWiT for Ad-Free Podcasts! Support what you love and get ad-free shows, a members-only Discord, and behind-the-scenes access. Join today: https://twit.tv/clubtwit Sponsors: 1password.com/securitynow zscaler.com/security bigid.com/securitynow uscloud.com