The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Sam Charrington

  • 42 minutes 9 seconds
    Controlling Fusion Reactor Instability with Deep Reinforcement Learning with Aza Jalalvand - #682

    Today we're joined by Azarakhsh (Aza) Jalalvand, a research scholar at Princeton University, to discuss his work using deep reinforcement learning to control plasma instabilities in nuclear fusion reactors. Aza explains his team developed a model to detect and avoid a fatal plasma instability called ‘tearing mode’. Aza walks us through the process of collecting and pre-processing the complex diagnostic data from fusion experiments, training the models, and deploying the controller algorithm on the DIII-D fusion research reactor. He shares insights from developing the controller and discusses the future challenges and opportunities for AI in enabling stable and efficient fusion energy production.


    The complete show notes for this episode can be found at twimlai.com/go/682.

    29 April 2024, 8:22 pm
  • 47 minutes 8 seconds
    GraphRAG: Knowledge Graphs for AI Applications with Kirk Marple - #681

    Today we're joined by Kirk Marple, CEO and founder of Graphlit, to explore the emerging paradigm of "GraphRAG," or Graph Retrieval Augmented Generation. In our conversation, Kirk digs into the GraphRAG architecture and how Graphlit uses it to offer a multi-stage workflow for ingesting, processing, retrieving, and generating content using LLMs (like GPT-4) and other Generative AI tech. He shares how the system performs entity extraction to build a knowledge graph and how graph, vector, and object storage are integrated in the system. We dive into how the system uses “prompt compilation” to improve the results it gets from Large Language Models during generation. We conclude by discussing several use cases the approach supports, as well as future agent-based applications it enables.


    The complete show notes for this episode can be found at twimlai.com/go/681.

    22 April 2024, 6:58 pm
  • 46 minutes 24 seconds
    Teaching Large Language Models to Reason with Reinforcement Learning with Alex Havrilla - #680

    Today we're joined by Alex Havrilla, a PhD student at Georgia Tech, to discuss "Teaching Large Language Models to Reason with Reinforcement Learning." Alex discusses the role of creativity and exploration in problem solving and explores the opportunities presented by applying reinforcement learning algorithms to the challenge of improving reasoning in large language models. Alex also shares his research on the effect of noise on language model training, highlighting the robustness of LLM architecture. Finally, we delve into the future of RL, and the potential of combining language models with traditional methods to achieve more robust AI reasoning.


    The complete show notes for this episode can be found at twimlai.com/go/680.

    16 April 2024, 10:58 pm
  • 49 minutes 46 seconds
    Localizing and Editing Knowledge in LLMs with Peter Hase - #679

    Today we're joined by Peter Hase, a fifth-year PhD student at the University of North Carolina NLP lab. We discuss "scalable oversight", and the importance of developing a deeper understanding of how large neural networks make decisions. We learn how matrices are probed by interpretability researchers, and explore the two schools of thought regarding how LLMs store knowledge. Finally, we discuss the importance of deleting sensitive information from model weights, and how "easy-to-hard generalization" could increase the risk of releasing open-source foundation models.


    The complete show notes for this episode can be found at twimlai.com/go/679.

    8 April 2024, 9:03 pm
  • 48 minutes 27 seconds
    Coercing LLMs to Do and Reveal (Almost) Anything with Jonas Geiping - #678

    Today we're joined by Jonas Geiping, a research group leader at the ELLIS Institute, to explore his paper: "Coercing LLMs to Do and Reveal (Almost) Anything". Jonas explains how neural networks can be exploited, highlighting the risk of deploying LLM agents that interact with the real world. We discuss the role of open models in enabling security research, the challenges of optimizing over certain constraints, and the ongoing difficulties in achieving robustness in neural networks. Finally, we delve into the future of AI security, and the need for a better approach to mitigate the risks posed by optimized adversarial attacks.


    The complete show notes for this episode can be found at twimlai.com/go/678.

    1 April 2024, 7:15 pm
  • 47 minutes 47 seconds
    V-JEPA, AI Reasoning from a Non-Generative Architecture with Mido Assran - #677

    Today we’re joined by Mido Assran, a research scientist at Meta’s Fundamental AI Research (FAIR). In this conversation, we discuss V-JEPA, a new model being billed as “the next step in Yann LeCun's vision” for true artificial reasoning. V-JEPA, the video version of Meta’s Joint Embedding Predictive Architecture, aims to bridge the gap between human and machine intelligence by training models to learn abstract concepts in a more efficient predictive manner than generative models. V-JEPA uses a novel self-supervised training approach that allows it to learn from unlabeled video data without being distracted by pixel-level detail. Mido walks us through the process of developing the architecture and explains why it has the potential to revolutionize AI.


    The complete show notes for this episode can be found at twimlai.com/go/677.

    25 March 2024, 4:00 pm
  • 49 minutes 34 seconds
    Video as a Universal Interface for AI Reasoning with Sherry Yang - #676

    Today we’re joined by Sherry Yang, senior research scientist at Google DeepMind and a PhD student at UC Berkeley. In this interview, we discuss her new paper, "Video as the New Language for Real-World Decision Making,” which explores how generative video models can play a role similar to language models as a way to solve tasks in the real world. Sherry draws the analogy between natural language as a unified representation of information and text prediction as a common task interface and demonstrates how video as a medium and generative video as a task exhibit similar properties. This formulation enables video generation models to play a variety of real-world roles as planners, agents, compute engines, and environment simulators. Finally, we explore UniSim, an interactive demo of Sherry's work and a preview of her vision for interacting with AI-generated environments.


    The complete show notes for this episode can be found at twimlai.com/go/676.

    18 March 2024, 5:09 pm
  • 40 minutes 26 seconds
    Assessing the Risks of Open AI Models with Sayash Kapoor - #675

    Today we’re joined by Sayash Kapoor, a Ph.D. student in the Department of Computer Science at Princeton University. Sayash walks us through his paper: "On the Societal Impact of Open Foundation Models.” We dig into the controversy around AI safety, the risks and benefits of releasing open model weights, and how we can establish common ground for assessing the threats posed by AI. We discuss the application of the framework presented in the paper to specific risks, such as the biosecurity risk of open LLMs, as well as the growing problem of "Non Consensual Intimate Imagery" using open diffusion models.


    The complete show notes for this episode can be found at twimlai.com/go/675.

    11 March 2024, 6:09 pm
  • 32 minutes 12 seconds
    OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia - #674

    Today we’re joined by Akshita Bhagia, a senior research engineer at the Allen Institute for AI. Akshita joins us to discuss OLMo, a new open source language model with 7 billion and 1 billion variants, but with a key difference compared to similar models offered by Meta, Mistral, and others. Namely, the fact that AI2 has also published the dataset and key tools used to train the model. In our chat with Akshita, we dig into the OLMo models and the various projects falling under the OLMo umbrella, including Dolma, an open three-trillion-token corpus for language model pretraining, and Paloma, a benchmark and tooling for evaluating language model performance across a variety of domains.


    The complete show notes for this episode can be found at twimlai.com/go/674.

    4 March 2024, 8:10 pm
  • 25 minutes 3 seconds
    Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski - #673

    Today we’re joined by Ben Prystawski, a PhD student in the Department of Psychology at Stanford University working at the intersection of cognitive science and machine learning. Our conversation centers on Ben’s recent paper, “Why think step by step? Reasoning emerges from the locality of experience,” which he recently presented at NeurIPS 2023. In this conversation, we start out exploring basic questions about LLM reasoning, including whether it exists, how we can define it, and how techniques like chain-of-thought reasoning appear to strengthen it. We then dig into the details of Ben’s paper, which aims to understand why thinking step-by-step is effective and demonstrates that local structure is the key property of LLM training data that enables it.


    The complete show notes for this episode can be found at twimlai.com/go/673.

    26 February 2024, 7:17 pm
  • 45 minutes 38 seconds
    Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh - #672

    Today we're joined by Armineh Nourbakhsh of JP Morgan AI Research to discuss the development and capabilities of DocLLM, a layout-aware large language model for multimodal document understanding. Armineh provides a historical overview of the challenges of document AI and an introduction to the DocLLM model. Armineh explains how this model, distinct from both traditional LLMs and document AI models, incorporates both textual semantics and spatial layout in processing enterprise documents like reports and complex contracts. We dig into her team’s approach to training DocLLM, their choice of a generative model as opposed to an encoder-based approach, the datasets they used to build the model, their approach to incorporating layout information, and the various ways they evaluated the model’s performance.


    The complete show notes for this episode can be found at twimlai.com/go/672.

    19 February 2024, 7:07 pm
  • More Episodes? Get the App
© MoonFM 2024. All rights reserved.