A podcast on Bayesian inference - the methods, the projects and the people who make it possible!
• Support & get perks!
• Bayesian Modeling course (first 2 lessons free)
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work !
Takeaways:
Q: Is generosity a natural human trait?
A: Yes, generosity is hardwired in our brains and is essential for social interaction.
Q: Why do people say they care about causes but not act on it?
A: There is often a disconnect between stated care for causes and actual action. Understanding the conditions under which generosity aligns with a person's identity is crucial for bridging this gap.
Q: How should fundraising efforts be approached?
A: Fundraising should primarily focus on belief updating rather than mere persuasion.
Q: What are the benefits of being generous?
A: Generosity has significant mental and physical health benefits, as the brain's reward systems activate when we give, making us feel good.
Q: How do our beliefs relate to our actions?
A: Our beliefs about ourselves strongly influence our actions and decisions, including our decision to be generous.
Q: Can generosity impact a community?
A: Yes, generosity can be a powerful tool for improving community dynamics.
Q: How can technology like AI assist institutions with donors?
A: AI could help institutions remember donors better, improving the donor-institution relationship.
00:00 What's the role of Behavioral Science inPhilanthropy
19:57 What is The Neuroscience of Generosity?
24:40 How can we best understand Donor Decision-Making?
32:14 How can we achieve reframe Beliefs and Actions?
35:39 What is the role of Identity in Habit Formation?
38:06 What is the Generosity Gap in Philanthropy?
45:06 How can we reduce Friction in Donation Processes?
48:27 What is the role of AI and Trust in Nonprofits?
52:11 How can we build Predictive Models for Donor Behavior?
55:41 What is the role of Empathy in Sales and Stakeholder Engagement?
01:00:46 How can we best align ideas with Stakeholder Beliefs?
01:02:06 How can we explore Generosity and Memory?
Thank you to my Patrons for making this episode possible!
Today's clip is from Episode 152 of the podcast, with Daniel Saunders.
In this conversation, Daniel Saunders explains how to incorporate risk aversion into Bayesian price optimization. The key insight is that uncertainty around expected profit is asymmetric across price points, low prices yield more predictable (if modest) returns, while high prices introduce much wider uncertainty. Rather than simply maximizing expected profit, you can pass profit through an exponential utility function that models diminishing returns, a well-established idea from economics.
This adds an adjustable risk aversion parameter to the optimization: as risk aversion increases, the model shifts toward more conservative price recommendations, trading off potentially large but uncertain gains for outcomes with tighter, more reliable distributions.
Get the full discussion here
• Join this channel to get access to perks:
https://www.patreon.com/c/learnbayesstats
• Intro to Bayes Course (first 2 lessons free): https://topmate.io/alex_andorra/503302
• Advanced Regression Course (first 2 lessons free): https://topmate.io/alex_andorra/1011122
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !
• Support & get perks!
• Proudly sponsored by PyMC Labs! Get in touch at [email protected]
• Intro to Bayes and Advanced Regression courses (first 2 lessons free)
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work !
00:00 The Importance of Decision-Making in Data Science
06:41 From Philosophy to Bayesian Statistics
14:57 The Role of Soft Skills in Data Science
18:19 Understanding Decision Theory Workflows
22:43 Shifting Focus from Accuracy to Business Value
26:23 Leveraging PyTensor for Optimization
34:27 Applying Optimal Decision-Making in Industry
40:06 Understanding Utility Functions in Regulation
41:35 Introduction to Obeisance Decision Theory Workflow
42:33 Exploring Price Elasticity and Demand
45:54 Optimizing Profit through Bayesian Models
51:12 Risk Aversion and Utility Functions
57:18 Advanced Risk Management Techniques
01:01:08 Practical Applications of Bayesian Decision-Making
01:06:54 Future Directions in Bayesian Inference
01:10:16 The Quest for Better Inference Algorithms
01:15:01 Dinner with a Polymath: Herbert Simon
Thank you to my Patrons for making this episode possible!
Today's clip is from Episode 151 of the podcast, with Jonas Arruda
In this conversation, Jonas Arruda explains how diffusion models generate data by learning to reverse a noise process. The idea is to start from a simple distribution like Gaussian noise and gradually remove noise until the target distribution emerges. This is done through a forward process that adds noise to clean parameters and a backward process that learns how to undo that corruption. A noise schedule controls how much noise is added or removed at each step, guiding the transformation from pure randomness back to meaningful structure.
Get the full discussion here
• Join this channel to get access to perks:
https://www.patreon.com/c/learnbayesstats
• Intro to Bayes Course (first 2 lessons free): https://topmate.io/alex_andorra/503302
• Advanced Regression Course (first 2 lessons free): https://topmate.io/alex_andorra/1011122
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !
• Support & get perks!
• Proudly sponsored by PyMC Labs! Get in touch at [email protected]
• Intro to Bayes and Advanced Regression courses (first 2 lessons free)
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work !
Chapters:
00:00 Exploring Generative AI and Scientific Modeling
10:27 Understanding Simulation-Based Inference (SBI) and Its Applications
15:59 Diffusion Models in Simulation-Based Inference
19:22 Live Coding Session: Implementing Baseflow for SBI
34:39 Analyzing Results and Diagnostics in Simulation-Based Inference
46:18 Hierarchical Models and Amortized Bayesian Inference
48:14 Understanding Simulation-Based Inference (SBI) and Its Importance
49:14 Diving into Diffusion Models: Basics and Mechanisms
50:38 Forward and Backward Processes in Diffusion Models
53:03 Learning the Score: Training Diffusion Models
54:57 Inference with Diffusion Models: The Reverse Process
57:36 Exploring Variants: Flow Matching and Consistency Models
01:01:43 Benchmarking Different Models for Simulation-Based Inference
01:06:41 Hierarchical Models and Their Applications in Inference
01:14:25 Intervening in the Inference Process: Adding Constraints
01:25:35 Summary of Key Concepts and Future Directions
Thank you to my Patrons for making this episode possible!
Links from the show:
- Come meet Alex at the Field of Play Conference in Manchester, UK, March 27, 2026!
- Jonas's Diffusion for SBI Tutorial & Review (Paper & Code)
- The BayesFlow Library
- Jonas on LinkedIn
- Jonas on GitHub
- Further reading for more mathematical details: Holderrieth & Erives
- 150 Fast Bayesian Deep Learning, with David Rügamer, Emanuel Sommer & Jakob Robnik
- 107 Amortized Bayesian Inference with Deep Neural Networks, with Marvin Schmitt
• Support & get perks!
• Proudly sponsored by PyMC Labs! Get in touch at [email protected]
• Intro to Bayes and Advanced Regression courses (first 2 lessons free)
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work !
Chapters:
00:00 Scaling Bayesian Neural Networks
04:26 Origin Stories of the Researchers
09:46 Research Themes in Bayesian Neural Networks
12:05 Making Bayesian Neural Networks Fast
16:19 Microcanonical Langevin Sampler Explained
22:57 Bottlenecks in Scaling Bayesian Neural Networks
29:09 Practical Tools for Bayesian Neural Networks
36:48 Trade-offs in Computational Efficiency and Posterior Fidelity
40:13 Exploring High Dimensional Gaussians
43:03 Practical Applications of Bayesian Deep Ensembles
45:20 Comparing Bayesian Neural Networks with Standard Approaches
50:03 Identifying Real-World Applications for Bayesian Methods
57:44 Future of Bayesian Deep Learning at Scale
01:05:56 The Evolution of Bayesian Inference Packages
01:10:39 Vision for the Future of Bayesian Statistics
Thank you to my Patrons for making this episode possible!
Come meet Alex at the Field of Play Conference in Manchester, UK, March 27, 2026!
Links from the show:
David Rügamer:
* Website
* Google Scholar
* GitHub
Emanuel Sommer:
* Website
* GitHub
* Google Scholar
Jakob Robnik:
* Google Scholar
* GitHub
* Microcanonical Langevin paper
* LinkedIn
Today’s clip is from episode 149 of the podcast, with Alana Karen.
This conversation explores the evolving landscape of technology, particularly in Silicon Valley, focusing on the cultural shifts due to mass layoffs, the debate over remote work, and the impact of AI on job roles and priorities. The discussion highlights the importance of adapting to these changes and preparing for the future by developing complex skills that AI cannot easily replicate.
Get the full discussion here!
• Join this channel to get access to perks:
https://www.patreon.com/c/learnbayesstats
• Intro to Bayes Course (first 2 lessons free): https://topmate.io/alex_andorra/503302
• Advanced Regression Course (first 2 lessons free): https://topmate.io/alex_andorra/1011122
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !
• Support & get perks!
• Proudly sponsored by PyMC Labs! Get in touch at [email protected]
• Intro to Bayes and Advanced Regression courses (first 2 lessons free)
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work !
Chapters:
11:37 The Hard Tech Era
21:08 The Shift in Tech Work Culture
28:49 AI's Impact on Job Security and Work Dynamics
34:33 Adapting to AI: Skills for the Future
45:56 Understanding AI Models and Their Limitations
47:25 The Importance of Diversity in AI Development
54:34 Positioning Technical Talent for Job Security
57:58 Building Resilience in Uncertain Times
01:06:33 Recognizing Diverse Ambitions in Career Progression
01:12:51 The Role of Managers in Employee Retention
01:26:55 Solving Complex Problems with AI and Innovation
Thank you to my Patrons for making this episode possible!
Links from the show:
Today’s clip is from episode 148 of the podcast, with Scott Berry.
In this conversation, Alex and Scott discuss emphasizing the shift from frequentist to Bayesian approaches in clinical trials.
They highlight the limitations of traditional trial designs and the advantages of adaptive and platform trials, particularly in the context of COVID-19 treatment.
The discussion provides insights into the complexities of trial design and the innovative methodologies that are shaping the future of medical research.
Get the full discussion here!
• Join this channel to get access to perks: https://www.patreon.com/c/learnbayesstats
• Intro to Bayes Course (first 2 lessons free): https://topmate.io/alex_andorra/503302
• Advanced Regression Course (first 2 lessons free): https://topmate.io/alex_andorra/1011122
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !
• Support & get perks!
• Proudly sponsored by PyMC Labs. Get in touch and tell them you come from LBS!
• Intro to Bayes and Advanced Regression courses (first 2 lessons free)
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work !
Chapters:
13:16 Understanding Adaptive and Platform Trials
25:25 Real-World Applications and Innovations in Trials
34:11 Challenges in Implementing Bayesian Adaptive Trials
42:09 The Birth of a Simulation Tool
44:10 The Importance of Simulated Data
48:36 Lessons from High-Stakes Trials
52:53 Navigating Adaptive Trial Designs
56:55 Communicating Complexity to Stakeholders
01:02:29 The Future of Clinical Trials
01:10:24 Skills for the Next Generation of Statisticians
Thank you to my Patrons for making this episode possible!
Yusuke Saito, Avi Bryant, Giuliano Cruz, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Aubrey Clayton, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady, Kurt TeKolste, Marcus Nölke, Maggi Mackintosh, Grant Pezzolesi, Joshua Meehl, Javier Sabio, Kristian Higgins, Matt Rosinski, Luis Fonseca, Dante Gates, Matt Niccolls, Maksim Kuznecov, Michael Thomas, Luke Gorrie, Cory Kiser, Julio, Edvin Saveljev, Frederick Ayala, Jeffrey Powell, Gal Kampel, Adan Romero, Blake Walters, Jonathan Morgan, Francesco Madrisotti, Ivy Huang, Gary Clarke, Robert Flannery, Rasmus Hindström, Stefan, Corey Abshire, Mike Loncaric, Ronald Legere, Sergio Dolia, Michael Cao, Yiğit Aşık, Suyog Chandramouli, Guillaume Berthon, Avenicio Baca, Spencer Boucher, Krzysztof Lechowski, Danimal, Jácint Juhász, Sander and Philippe.
Links from the show:
Today’s clip is from episode 147 of the podcast, with Martin Ingram.
Alex and Martin discuss the intricacies of variational inference, particularly focusing on the ADVI method and its challenges. They explore the evolution of approximate inference methods, the significance of mean field variational inference, and the innovative linear response technique for covariance estimation.
The discussion also delves into the trade-offs between stochastic and deterministic optimization techniques, providing insights into their implications for Bayesian statistics.
Get the full discussion here.
Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work!
Visit our Patreon page to unlock exclusive Bayesian swag ;)
Transcript
This is an automatic transcript and may therefore contain errors. Please get in touch if you're willing to correct them.