The New Stack Podcast

The New Stack Podcast

The New Stack Podcast is all about the developers, software engineers and operations people who build at-scale architectures that change the way we develop and deploy software. Subscribe to TNS on YouTube at: https://www.youtube.com/c/TheNewStack

  • 23 minutes 29 seconds
    Kubernetes Gets Back to Scaling with Virtual Clusters

    A virtual cluster, described by Loft Labs CEO Lukas Gentele at Kubecon+ CloudNativeCon Paris, is a Kubernetes control plane running inside a container within another Kubernetes cluster. In this New Stack Makers episode, Gentele explained that this approach eliminates the need for numerous separate control planes, allowing VMs to run in lightweight, quickly deployable containers. Loft Labs' open-sourced vcluster technology enables virtual clusters to spin up in about six seconds, significantly faster than traditional Kubernetes clusters that can take over 30 minutes to start in services like Amazon EKS or Google GKE.

    The integration of vCluster into Rancher at KubeCon Paris enables users to manage virtual clusters alongside real clusters seamlessly. This innovation addresses challenges faced by companies managing multiple applications and clusters, advocating for a multi-tenant cluster approach for improved sharing and security, contrary to the trend of isolated single-tenant clusters that emerged due to complexities in cluster sharing within Kubernetes.

     

    Learn more from The New Stack about virtual clusters: 

    Vcluster to the Rescue 

    Navigating the Trade-Offs of Scaling Kubernetes Dev Environments 

    Managing Kubernetes Clusters for Platform Engineers 

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game. https://thenewstack.io/newsletter/

    25 April 2024, 11:00 am
  • 28 minutes 39 seconds
    How Giant Swarm Is Helping to Support the Future of Flux

    When Weaveworks, known for pioneering "GitOps," shut down, concerns arose about the future of Flux, a critical open-source project. However, Puja Abbassi, Giant Swarm's VP of Product, reassured Alex Williams, Founder and Publisher of The New Stack at Open Source Summit in Paris that Flux's maintenance is secure in this episode of The New Makers podcast. 

    Giant companies like Microsoft Azure and GitLab have pledged support. Giant Swarm, an avid Flux user, also contributes to its development, ensuring its vitality alongside related projects like infrastructure code plugins and UI improvements. Abbassi highlighted the importance of considering a project's sustainability and integration capabilities when choosing open-source tools. He noted Argo CD's advantage in UI, emphasizing that projects like Flux must evolve to meet user expectations and avoid being overshadowed. This underscores the crucial role of community support, diversity, and compatibility within the Cloud Native Computing Foundation's ecosystem for long-term tool adoption.

    Learn more from The New Stack about Flux:  

    End of an Era: Weaveworks Closes Shop Amid Cloud Native Turbulence 

    Why Flux Isn't Dying after Weaveworks

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game. 

     

     

    22 April 2024, 11:00 am
  • 37 minutes 31 seconds
    AI, LLMs and Security: How to Deal with the New Threats

    The use of large language models (LLMs) has become widespread, but there are significant security risks associated with them. LLMs with millions or billions of parameters are complex and challenging to fully scrutinize, making them susceptible to exploitation by attackers who can find loopholes or vulnerabilities. On an episode of The New Stack Makers, Chris Pirillo, Tech Evangelist and Lance Seidman, Backend Engineer at Atomic Form discussed these security challenges, emphasizing the need for human oversight to protect AI systems.

    One example highlighted was malicious AI models on Hugging Face, which exploited the Python pickle module to execute arbitrary commands on users' machines. To mitigate such risks, Hugging Face implemented security scanners to check every file for security threats. However, human vigilance remains crucial in identifying and addressing potential exploits.

    Seidman also stressed the importance of technical safeguards and a culture of security awareness within the AI community. Developers should prioritize security throughout the development life cycle to stay ahead of evolving threats. Overall, the message is clear: while AI offers remarkable capabilities, it requires careful management and oversight to prevent misuse and protect against security breaches.

    Learn more from The New Stack about AI and security:  

    Artificial Intelligence: Stopping the Big Unknown in Application, Data Security 

    Cyberattacks, AI and Multicloud Hit Cybersecurity in 2023

    Will Generative AI Kill DevSecOps? 

     

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

     

    11 April 2024, 11:00 am
  • 29 minutes 29 seconds
    How Kubernetes Faces a New Reality with the AI Engineer

    The Kubernetes community primarily focuses on improving the development and operations experience for applications and infrastructure, emphasizing DevOps and developer-centric approaches. In contrast, the data science community historically moved at a slower pace. However, with the emergence of the AI engineer persona, the pace of advancement in data science has accelerated significantly. 

    Alex Williams, founder and publisher of The New Stack co-hosted a discussion with Sanjeev Mohan, an independent analyst, which highlighted the challenges faced by data-related tasks on Kubernetes due to the stateful nature of data. Unlike applications, restarting a database node after a failure may lead to inconsistent states and data loss. This discrepancy in pace and needs between developers and data scientists led to Kubernetes and the Cloud Native Computing Foundation initially overlooking data science. 

    Nevertheless, Mohan noted that the pace of data engineers has increased as they explore new AI applications and workloads. Kubernetes now plays a crucial role in supporting these advancements by helping manage resources efficiently, especially considering the high cost of training large language models (LLMs) and using GPUs for AI workloads. Mohan also discussed the evolving landscape of AI frameworks and the importance of aligning business use cases with AI strategies. Learn more from The New Stack about data development and DevOps: AI Will Drive Streaming Data Use — But Not Yet, Report Says https://thenewstack.io/ai-will-drive-streaming-data-adoption-says-redpanda-survey/ The Paradigm Shift from Model-Centric to Data-Centric AI https://thenewstack.io/the-paradigm-shift-from-model-centric-to-data-centric-ai/ AI Development Needs to Focus More on Data, Less on Models https://thenewstack.io/ai-development-needs-to-focus-more-on-data-less-on-models/ 

    Learn more from The New Stack about data development and DevOps: 

    AI Will Drive Streaming Data Use - But Not Yet, Report Says

    The Paradigm Shift from Model-Centric to Data-Centric AI

    AI Development Needs to Focus More on Data, Less on Models 

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game. https://thenewstack.io/newsletter/

    4 April 2024, 11:00 am
  • 25 minutes 51 seconds
    LLM Observability: The Breakdown

    LLM observability focuses on maximizing the utility of larger language models (LLMs) by monitoring key metrics and signals. Alex Williams, Founder and Publisher for The New Stack, and Janikiram MSV, Principal of Janikiram & Associates and an analyst and writer for The New Stack, discusses the emergence of the LLM stack, which encompasses various components like LLMs, vector databases, embedding models, retrieval systems, read anchor models, and more. The objective of LLM observability is to ensure that users can extract desired outcomes effectively from this complex ecosystem.

    Similar to infrastructure observability in DevOps and SRE practices, LLM observability aims to provide insights into the LLM stack's performance. This includes monitoring metrics specific to LLMs, such as GPU/CPU usage, storage, model serving, change agents in applications, hallucinations, span traces, relevance, retrieval models, latency, monitoring, and user feedback. MSV emphasizes the importance of monitoring resource usage, model catalog synchronization with external providers like Hugging Face, vector database availability, and the inference engine's functionality.

    He also mentions peer companies in the LLM observability space like Datadog, New Relic, Signoz, Dynatrace, LangChain (LangSmith), Arize.ai (Phoenix), and Truera, hinting at a deeper exploration in a future episode of The New Stack Makers.

     

    Learn more from The New Stack about LLM and observability  

    Observability in 2024: More OpenTelemetry, Less Confusion 

    How AI Can Supercharge Observability 

    Next-Gen Observability: Monitoring and Analytics in Platform Engineering

     

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game. 

     

     

    28 March 2024, 2:45 pm
  • 38 minutes 55 seconds
    Why Software Developers Should Be Thinking About the Climate

    In a conversation on The New Stack Makers, co-hosted by Alex Williams, TNS founder and publisher, and Charles Humble, an industry expert who served as a software engineer, architect and CTO and now podcaster, author and consultant at Conissaunce Ltd., discussed why software developers and engineers should care about their impact on climate change. Humble emphasized that building software sustainably starts with better operations, leading to cost savings and improved security. He cited past successes in combating environmental issues like acid rain and the ozone hole through international agreements and emissions reduction strategies.

    Despite modest growth since 2010, data centers remain significant electricity consumers, comparable to countries like Brazil. The power-intensive nature of AI models exacerbates these challenges and may lead to scarcity issues. Humble mentioned the Green Software Foundation's Maturity Matrix with goals for carbon-free data centers and longer device lifespans, discussing their validity and the role of regulation in achieving them. Overall, software development's environmental impact, primarily carbon emissions, necessitates proactive measures and industry-wide collaboration.

     

    Learn more from The New Stack about sustainability: 

    What is GreenOps? Putting a Sustainable Focus on FinOps

    Unraveling the Costs of Bad Code in Software Development 

    Can Reducing Cloud Waste Help Save the Planet?

    How to Build Open Source Sustainability

     

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    21 March 2024, 11:00 am
  • 39 minutes 45 seconds
    Nvidia’s Superchips for AI: ‘Radical,’ but a Work in Progress

    This New Stack Makers podcast co-hosted by Alex Williams, TNS founder and publisher, and Adrian Cockcroft, Partner and Analyst at OrionX.net, discussed Nvidia's GH200 Grace Hopper superchip. Industry expert Sunil Mallya, Co-founder and CTO of Flip AI weighed in on how it is revolutionizing the hardware industry for AI workloads by centralizing GPU communication, reducing networking overhead, and creating a more efficient system. 

    Mallya noted that despite its innovative design, challenges remain in adoption due to interface issues and the need for software to catch up with hardware advancements. However, optimism persists for the future of AI-focused chips, with Nvidia leading the charge in creating large-scale coherent memory systems. Meanwhile, Flip AI, a DevOps large language model, aims to interpret observability data to troubleshoot incidents effectively across various cloud platforms. While discussing the latest chip innovations and challenges in training large language models, the episode sheds light on the evolving landscape of AI hardware and software integration.

    Learn more from The New Stack about Nvidia and the future of chip design  

    Nvidia Wants to Rewrite the Software Development Stack 

    Nvidia GPU Dominance at a Crossroads 

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game.  

    14 March 2024, 11:00 am
  • 29 minutes 34 seconds
    Is GitHub Copilot Dependable? These Demos Aren’t Promising

    This New Stack Makers podcast co-hosted by TNS founder and publisher, Alex Williams and Joan Westenberg, founder and writer of Joan’s Index, discussed Copilot. Westenberg highlighted its integration with Microsoft 365 and its role as a coding assistant, showcasing its potential to streamline various tasks. 

    However, she also revealed its limitations, particularly in reliability. Despite being designed to assist with tasks across Microsoft 365, Copilot's performance fell short during Westenberg's tests, failing to retrieve necessary information from her email and Microsoft Teams meetings. While Copilot proves useful for coding, providing helpful code snippets, its effectiveness diminishes for more complex projects. Westenberg's demonstrations underscored both the strengths and weaknesses of Copilot, emphasizing the need for improvement, especially in reliability, to fulfill its promise as a versatile work companion.

     

    Learn more from The New Stack about Copilot 

    Microsoft One-ups Google with Copilot Stack for Developers 

    Copilot Enterprises Introduces Search and Customized Best Practices 

     

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game

     

    7 March 2024, 12:00 pm
  • 27 minutes 3 seconds
    The New Monitoring for Services That Feed from LLMs

    This New Stack Makers podcast co-hosted by Adrian Cockroft, analyst at OrionX.net and TNS founder and publisher, Alex Williams discusses the importance of monitoring services utilizing Large Language Models (LLMs) and the emergence of tools like LangChain and LangSmith to address this need. Adrian Cockcroft, formerly of Netflix and now working with The New Stack, highlights the significance of monitoring AI apps using LLMs and the challenges posed by slow and expensive API calls from LLMs. LangChain acts as middleware, connecting LLMs with services, akin to the Java Database Controller. LangChain's monitoring capabilities led to the development of LangSmith, a monitoring tool. Another tool, LangKit by WhyLabs, offers similar functionalities but is less integrated. This reflects the typical evolution of open-source projects into commercial products. LangChain recently secured funding, indicating growing interest in such monitoring solutions. Cockcroft emphasizes the importance of enterprise-level support and tooling for integrating these solutions into commercial environments. This discussion  underscores the evolving landscape of monitoring services powered by LLMs and the emergence of specialized tools to address associated challenges.

     

    Learn more from The New Stack about LangChain: 

    LangChain: The Trendiest Web Framework of 2023, Thanks to AI 

    How Retool AI Differs from LangChain (Hint: It's Automation) 

     

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game

     

    28 February 2024, 12:00 pm
  • 18 minutes 52 seconds
    How Platform Engineering Supports SRE

    In this New Stack Makers podcast, Martin Parker, a solutions architect for UST, spoke with TNS editor-in-chief, Heather Joslyn and discussed the significance of internal developer platforms (IDPs), emphasizing benefits beyond frontend developers to backend engineers and site reliability engineers (SREs). 

    Parker highlighted the role of IDPs in automating repetitive tasks, allowing SREs to focus on optimizing application performance. Standardization is key, ensuring observability and monitoring solutions align with best practices and cater to SRE needs. By providing standardized service level indicators (SLIs) and key performance indicators (KPIs), IDPs enable SREs to maintain reliability efficiently. Parker stresses the importance of avoiding siloed solutions by establishing standardized practices and tools for effective monitoring and incident response. Overall, the deployment of IDPs aims to streamline operations, reduce incidents, and enhance organizational value by empowering SREs to concentrate on system maintenance and improvements.

    Learn more from The New Stack about UST: 

    Cloud Cost-Unit Economics- A Modern Profitability Model 

    Cloud Native Users Struggle to Achieve Benefits, Report Says 

    John our community of newsletter subscribers to stay on top of the news and at the top of your game

    7 February 2024, 12:00 pm
  • 15 minutes 23 seconds
    Internal Developer Platforms: Helping Teams Limit Scope

    In this New Stack Makers podcast, Ben Wilcock, a senior technical marketing architect for Tanzu, spoke with TNS editor-in-chief, Heather Joslyn and discussed the challenges organizations face when building internal developer platforms, particularly the issue of scope, at KubeCon + CloudNativeCon North America. 

    He emphasized the difficulty for platform engineering teams to select and integrate various Kubernetes projects amid a plethora of options. Wilcock highlights the complexity of tracking software updates, new features, and dependencies once choices are made. He underscores the advantage of having a standardized approach to software deployment, preventing errors caused by diverse mechanisms. 

    Tanzu aims to simplify the adoption of platform engineering and internal developer platforms, offering a turnkey approach with the Tanzu Application Platform. This platform is designed to be flexible, malleable, and functional out of the box. Additionally, Tanzu has introduced the Tanzu Developer Portal, providing a focal point for developers to share information and facilitating faster progress in platform engineering without the need to integrate numerous open source projects.

     

    Learn more from The New Stack about Tanzu and internal developer platforms:

    VMware Unveils a Pile of New Data Services for Its Cloud VMware 

    VMware Expands Tanzu into a Full Platform Engineering Environment 

    VMware Targets the Platform Engineer

     

    Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

     

    31 January 2024, 12:00 pm
  • More Episodes? Get the App
© MoonFM 2024. All rights reserved.