LessWrong Curated Podcast

LessWrong

Audio version of the posts shared in the LessWrong Curated newsletter.

  • 5 minutes 50 seconds
    "Backyard cat fight shows Schelling points preexist language" by jchan
    Two cats fighting for control over my backyard appear to have settled on a particular chain-link fence as the delineation between their territories. This suggests that:

    1. Animals are capable of recognizing Schelling points
    2. Therefore, Schelling points do not depend on language for their Schelling-ness
    3. Therefore, tacit bargaining should be understood not as a special case of bargaining where communication happens to be restricted, but rather as the norm from which the exceptional case of explicit bargaining is derived.
    Summary of cat situation

    I don't have any pets, so my backyard is terra nullius according to Cat Law. This situation is unstable, as there are several outdoor cats in the neighborhood who would like to claim it. Our two contenders are Tabby Cat, who lives on the other side of the waist-high chain-link fence marking the back edge of my lot, and Tuxedo Cat, who lives in the place next-door to me.

    | || Tabby's || yard || (A) |------+...........+--------| (B) || | Tuxedo's| My yard | yard| ||| -- tall wooden fences.... short chain-link fence In the first [...]

    ---

    Outline:

    (00:43) Summary of cat situation

    (03:28) Why the fence?

    (04:00) If animals have Schelling points, then...

    ---

    First published:
    January 14th, 2026

    Source:
    https://www.lesswrong.com/posts/uYr8pba7TqaPpszX5/backyard-cat-fight-shows-schelling-points-preexist-language

    ---



    Narrated by TYPE III AUDIO.

    16 January 2026, 10:30 pm
  • 37 minutes 44 seconds
    "How AI Is Learning to Think in Secret" by Nicholas Andresen
    On Thinkish, Neuralese, and the End of Readable Reasoning

    In September 2025, researchers published the internal monologue of OpenAI's GPT-o3 as it decided to lie about scientific data. This is what it thought:

    Pardon? This looks like someone had a stroke during a meeting they didn’t want to be in, but their hand kept taking notes.

    That transcript comes from a recent paper published by researchers at Apollo Research and OpenAI on catching AI systems scheming. To understand what's happening here - and why one of the most sophisticated AI systems in the world is babbling about “synergy customizing illusions” - it first helps to know how we ended up being able to read AI thinking in the first place.

    That story starts, of all places, on 4chan.

    In late 2020, anonymous posters on 4chan started describing a prompting trick that would change the course of AI development. It was almost embarrassingly simple: instead of just asking GPT-3 for an answer, ask it instead to show its work before giving its final answer.

    Suddenly, it started solving math problems that had stumped it moments before.

    To see why, try multiplying 8,734 × 6,892 in your head. If you’re like [...]

    The original text contained 3 footnotes which were omitted from this narration.

    ---

    First published:
    January 6th, 2026

    Source:
    https://www.lesswrong.com/posts/gpyqWzWYADWmLYLeX/how-ai-is-learning-to-think-in-secret

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Robot thinking about alien symbols while paperclip character offers help with language development.Screenshot showing repetitive, incoherent text about disclaimers, illusions, and overshadowing.Diagram showing user input, AI reasoning about deception, and tool call manipulating data.Children's book page showing Frog and Toad with a cookie box.
    9 January 2026, 4:15 am
  • 5 minutes 37 seconds
    "On Owning Galaxies" by Simon Lermen
    It seems to be a real view held by serious people that your OpenAI shares will soon be tradable for moons and galaxies. This includes eminent thinkers like Dwarkesh Patel, Leopold Aschenbrenner, perhaps Scott Alexander and many more. According to them, property rights will survive an AI singularity event and soon economic growth is going to make it possible for individuals to own entire galaxies in exchange for some AI stocks. It follows that we should now seriously think through how we can equally distribute those galaxies and make sure that most humans will not end up as the UBI underclass owning mere continents or major planets.

    I don't think this is a particularly intelligent view. It comes from a huge lack of imagination for the future.

    Property rights are weird, but humanity dying isn't

    People may think that AI causing human extinction is something really strange and specific to happen. But it's the opposite: humans existing is a very brittle and strange state of affairs. Many specific things have to be true for us to be here, and when we build ASI there are many preferences and goals that would see us wiped out. It's actually hard to [...]

    ---

    Outline:

    (01:06) Property rights are weird, but humanity dying isnt

    (01:57) Why property rights wont survive

    (03:10) Property rights arent enough

    (03:36) What if there are many unaligned AIs?

    (04:18) Why would they be rewarded?

    (04:48) Conclusion

    ---

    First published:
    January 6th, 2026

    Source:
    https://www.lesswrong.com/posts/SYyBB23G3yF2v59i8/on-owning-galaxies

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Political cartoon showing person holding OpenAI stock certificate as AI takeover news plays on TV with nanobots swirling around.Board meeting with executives, AI system, CEO, and screen displayingApple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    8 January 2026, 5:15 pm
  • 50 minutes 46 seconds
    "AI Futures Timelines and Takeoff Model: Dec 2025 Update" by elifland, bhalstead, Alex Kastner, Daniel Kokotajlo
    We’ve significantly upgraded our timelines and takeoff models! It predicts when AIs will reach key capability milestones: for example, Automated Coder / AC (full automation of coding) and superintelligence / ASI (much better than the best humans at virtually all cognitive tasks). This post will briefly explain how the model works, present our timelines and takeoff forecasts, and compare it to our previous (AI 2027) models (spoiler: the AI Futures Model predicts about 3 years longer timelines to full coding automation than our previous model, mostly due to being less bullish on pre-full-automation AI R&D speedups).

    If you’re interested in playing with the model yourself, the best way to do so is via this interactive website: aifuturesmodel.com

    If you’d like to skip the motivation for our model to an explanation for how it works, go here, The website has a more in-depth explanation of the model (starts here; use the diagram on the right as a table of contents), as well as our forecasts.

    Why do timelines and takeoff modeling?

    The future is very hard to predict. We don't think this model, or any other model, should be trusted completely. The model takes into account what we think are [...]

    ---

    Outline:

    (01:32) Why do timelines and takeoff modeling?

    (03:18) Why our approach to modeling? Comparing to other approaches

    (03:24) AGI  timelines forecasting methods

    (03:29) Trust the experts

    (04:35) Intuition informed by arguments

    (06:10) Revenue extrapolation

    (07:15) Compute extrapolation anchored by the brain

    (09:53) Capability benchmark trend extrapolation

    (11:44) Post-AGI takeoff forecasts

    (13:33) How our model works

    (14:37) Stage 1: Automating coding

    (16:54) Stage 2: Automating research taste

    (18:18) Stage 3: The intelligence explosion

    (20:35) Timelines and takeoff forecasts

    (21:04) Eli

    (24:34) Daniel

    (38:32) Comparison to our previous (AI 2027) timelines and takeoff models

    (38:49) Timelines to Superhuman Coder (SC)

    (43:33) Takeoff from Superhuman Coder onward

    The original text contained 31 footnotes which were omitted from this narration.

    ---

    First published:
    December 31st, 2025

    Source:
    https://www.lesswrong.com/posts/YABG5JmztGGPwNFq2/ai-futures-timelines-and-takeoff-model-dec-2025-update

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Interactive webpage showing AI development timelines, forecasts, and efficiency graphs with adjustable parameters.Grap</truncato-artificial-root>
    6 January 2026, 12:15 pm
  • 13 minutes 51 seconds
    "In My Misanthropy Era" by jenn
    For the past year I've been sinking into the Great Books via the Penguin Great Ideas series, because I wanted to be conversant in the Great Conversation. I am occasionally frustrated by this endeavour, but overall, it's been fun! I'm learning a lot about my civilization and the various curmudgeons that shaped it.

    But one dismaying side effect is that it's also been quite empowering for my inner 13 year old edgelord. Did you know that before we invented woke, you were just allowed to be openly contemptuous of people?

    Here's Schopenhauer on the common man:

    They take an objective interest in nothing whatever. Their attention, not to speak of their mind, is engaged by nothing that does not bear some relation, or at least some possible relation, to their own person: otherwise their interest is not aroused. They are not noticeably stimulated even by wit or humour; they hate rather everything that demands the slightest thought. Coarse buffooneries at most excite them to laughter: apart from that they are earnest brutes – and all because they are capable of only subjective interest. It is precisely this which makes card-playing the most appropriate amusement for them – card-playing for [...]

    The original text contained 3 footnotes which were omitted from this narration.

    ---

    First published:
    January 4th, 2026

    Source:
    https://www.lesswrong.com/posts/otgrxjbWLsrDjbC2w/in-my-misanthropy-era

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Clown makeup meme about philosophy meetups and grad students.Document outlining four core values: Curiosity, Diversity of Perspectives, Inclusion, and Humility.Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    5 January 2026, 4:30 pm
  • 21 minutes 53 seconds
    "2025 in AI predictions" by jessicata
    Past years: 2023 2024

    Continuing a yearly tradition, I evaluate AI predictions from past years, and collect a convenience sample of AI predictions made this year. In terms of selection, I prefer selecting specific predictions, especially ones made about the near term, enabling faster evaluation.

    Evaluated predictions made about 2025 in 2023, 2024, or 2025 mostly overestimate AI capabilities advances, although there's of course a selection effect (people making notable predictions about the near-term are more likely to believe AI will be impressive near-term).

    As time goes on, "AGI" becomes a less useful term, so operationalizing predictions is especially important. In terms of predictions made in 2025, there is a significant cluster of people predicting very large AI effects by 2030. Observations in the coming years will disambiguate.

    Predictions about 2025

    2023

    Jessica Taylor: "Wouldn't be surprised if this exact prompt got solved, but probably something nearby that's easy for humans won't be solved?"

    The prompt: "Find a sequence of words that is: - 20 words long - contains exactly 2 repetitions of the same word twice in a row - contains exactly 2 repetitions of the same word thrice in a row"

    Self-evaluation: False; I underestimated LLM progress [...]

    ---

    Outline:

    (01:09) Predictions about 2025

    (03:35) Predictions made in 2025 about 2025

    (05:31) Predictions made in 2025

    ---

    First published:
    January 1st, 2026

    Source:
    https://www.lesswrong.com/posts/69qnNx8S7wkSKXJFY/2025-in-ai-predictions

    ---



    Narrated by TYPE III AUDIO.

    3 January 2026, 1:15 am
  • 17 minutes 47 seconds
    "Good if make prior after data instead of before" by dynomight
    They say you’re supposed to choose your prior in advance. That's why it's called a “prior”. First, you’re supposed to say say how plausible different things are, and then you update your beliefs based on what you see in the world.

    For example, currently you are—I assume—trying to decide if you should stop reading this post and do something else with your life. If you’ve read this blog before, then lurking somewhere in your mind is some prior for how often my posts are good. For the sake of argument, let's say you think 25% of my posts are funny and insightful and 75% are boring and worthless.

    OK. But now here you are reading these words. If they seem bad/good, then that raises the odds that this particular post is worthless/non-worthless. For the sake of argument again, say you find these words mildly promising, meaning that a good post is 1.5× more likely than a worthless post to contain words with this level of quality.

    If you combine those two assumptions, that implies that the probability that this particular post is good is 33.3%. That's true because the red rectangle below has half the area of the blue [...]

    ---

    Outline:

    (03:28) Aliens

    (07:06) More aliens

    (09:28) Huh?

    ---

    First published:
    December 18th, 2025

    Source:
    https://www.lesswrong.com/posts/JAA2cLFH7rLGNCeCo/good-if-make-prior-after-data-instead-of-before

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Diagram showing probability brackets P[good] in red and P[bad] in blue.Statistical diagram showing probability distributions for good words versus bad words.Diagram showing probability notation for good words versus bad words.Diagram showing probability notation for aliens and no aliens events.
    27 December 2025, 7:15 pm
  • 12 minutes 46 seconds
    "Measuring no CoT math time horizon (single forward pass)" by ryan_greenblatt
    A key risk factor for scheming (and misalignment more generally) is opaque reasoning ability.One proxy for this is how good AIs are at solving math problems immediately without any chain-of-thought (CoT) (as in, in a single forward pass).I've measured this on a dataset of easy math problems and used this to estimate 50% reliability no-CoT time horizon using the same methodology introduced in Measuring AI Ability to Complete Long Tasks (the METR time horizon paper).

    Important caveat: To get human completion times, I ask Opus 4.5 (with thinking) to estimate how long it would take the median AIME participant to complete a given problem.These times seem roughly reasonable to me, but getting some actual human baselines and using these to correct Opus 4.5's estimates would be better.

    Here are the 50% reliability time horizon results:

    I find that Opus 4.5 has a no-CoT 50% reliability time horizon of 3.5 minutes and that time horizon has been doubling every 9 months.

    In an earlier post (Recent LLMs can leverage filler tokens or repeated problems to improve (no-CoT) math performance), I found that repeating the problem substantially boosts performance.In the above plot, if [...]

    ---

    Outline:

    (02:33) Some details about the time horizon fit and data

    (04:17) Analysis

    (06:59) Appendix: scores for Gemini 3 Pro

    (11:41) Appendix: full result tables

    (11:46) Time Horizon - Repetitions

    (12:07) Time Horizon - Filler

    The original text contained 2 footnotes which were omitted from this narration.

    ---

    First published:
    December 26th, 2025

    Source:
    https://www.lesswrong.com/posts/Ty5Bmg7P6Tciy2uj2/measuring-no-cot-math-time-horizon-single-forward-pass

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Two graphs comparingTwo graphs comparing AI model training times with and without repeat/filler, linear and log scale.Graph showing
    27 December 2025, 6:45 am
  • 36 minutes 52 seconds
    "Recent LLMs can use filler tokens or problem repeats to improve (no-CoT) math performance" by ryan_greenblatt
    Prior results have shown that LLMs released before 2024 can't leverage 'filler tokens'—unrelated tokens prior to the model's final answer—to perform additional computation and improve performance.[1]I did an investigation on more recent models (e.g. Opus 4.5) and found that many recent LLMs improve substantially on math problems when given filler tokens.That is, I force the LLM to answer some math question immediately without being able to reason using Chain-of-Thought (CoT), but do give the LLM filter tokens (e.g., text like "Filler: 1 2 3 ...") before it has to answer.Giving Opus 4.5 filler tokens[2] boosts no-CoT performance from 45% to 51% (p=4e-7) on a dataset of relatively easy (competition) math problems[3].I find a similar effect from repeating the problem statement many times, e.g. Opus 4.5's no-CoT performance is boosted from 45% to 51%.

    Repeating the problem statement generally works better and is more reliable than filler tokens, especially for relatively weaker models I test (e.g. Qwen3 235B A22B), though the performance boost is often very similar, especially for Anthropic models.The first model that is measurably uplifted by repeats/filler is Opus 3, so this effect has been around for a [...]

    ---

    Outline:

    (03:59) Datasets

    (06:01) Prompt

    (06:58) Results

    (07:01) Performance vs number of repeats/filler

    (08:29) Alternative types of filler tokens

    (09:03) Minimal comparison on many models

    (10:05) Relative performance improvement vs default absolute performance

    (13:59) Comparing filler vs repeat

    (14:44) Future work

    (15:53) Appendix: How helpful were AIs for this project?

    (16:44) Appendix: more information about Easy-Comp-Math

    (25:05) Appendix: tables with exact values

    (25:11) Gen-Arithmetic - Repetitions

    (25:36) Gen-Arithmetic - Filler

    (25:59) Easy-Comp-Math - Repetitions

    (26:21) Easy-Comp-Math - Filler

    (26:45) Partial Sweep - Gen-Arithmetic - Repetitions

    (27:07) Partial Sweep - Gen-Arithmetic - Filler

    (27:27) Partial Sweep - Easy-Comp-Math - Repetitions

    (27:48) Partial Sweep - Easy-Comp-Math - Filler

    (28:09) Appendix: more plots

    (28:14) Relative accuracy improvement plots

    (29:05) Absolute accuracy improvement plots

    (29:57) Filler vs repeat comparison plots

    (30:02) Absolute

    (30:30) Relative

    (30:58) Appendix: significance matrices

    (31:03) Easy-Comp-Math - Repetitions

    (32:27) Easy-Comp-Math - Filler

    (33:50) Gen-Arithmetic - Repetitions

    (35:12) Gen-Arithmetic - Filler

    The original text contained 9 footnotes which were omitted from this narration.

    ---

    First published:
    December 22nd, 2025

    Source:
    https://www.lesswrong.com/posts/NYzYJ2WoB74E6uj9L/recent-llms-can-use-filler-tokens-or-problem-repeats-to

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Bar graph show</truncato-artificial-root>
    23 December 2025, 11:15 pm
  • 5 minutes 3 seconds
    "Turning 20 in the probable pre-apocalypse" by Parv Mahajan
    Master version of this on https://parvmahajan.com/2025/12/21/turning-20.html

    I turn 20 in January, and the world looks very strange. Probably, things will change very quickly. Maybe, one of those things is whether or not we’re still here.

    This moment seems very fragile, and perhaps more than most moments will never happen again. I want to capture a little bit of what it feels like to be alive right now.

    1. 

    Everywhere around me there is this incredible sense of freefall and of grasping. I realize with excitement and horror that over a semester Claude went from not understanding my homework to easily solving it, and I recognize this is the most normal things will ever be. Suddenly, the ceiling for what is possible seems so high - my classmates join startups, accelerate their degrees; I find myself building bespoke bioinformatics tools in minutes, running month-long projects in days. I write dozens of emails and thousands of lines of code a week, and for the first time I no longer feel limited by my ability but by my willpower. I spread the gospel to my friends - “there has never been a better time to have a problem” - even as [...]

    ---

    Outline:

    (00:34) 1.

    (03:12) 2.

    ---

    First published:
    December 21st, 2025

    Source:
    https://www.lesswrong.com/posts/S5dnLsmRbj2JkLWvf/turning-20-in-the-probable-pre-apocalypse

    ---



    Narrated by TYPE III AUDIO.

    23 December 2025, 8:45 pm
  • 20 minutes 57 seconds
    "Alignment Pretraining: AI Discourse Causes Self-Fulfilling (Mis)alignment" by Cam, Puria Radmard, Kyle O’Brien, David Africa, Samuel Ratnam, andyk
    TL;DR

    LLMs pretrained on data about misaligned AIs themselves become less aligned. Luckily, pretraining LLMs with synthetic data about good AIs helps them become more aligned. These alignment priors persist through post-training, providing alignment-in-depth. We recommend labs pretrain for alignment, just as they do for capabilities.

    Website: alignmentpretraining.ai
    Us: geodesicresearch.org | x.com/geodesresearch

    Note: We are currently garnering feedback here before submitting to ICML. Any suggestions here or on our Google Doc are welcome! We will be releasing a revision on arXiv in the coming days. Folks who leave feedback will be added to the Acknowledgment section. Thank you!

    Abstract

    We pretrained a suite of 6.9B-parameter LLMs, varying only the content related to AI systems, and evaluated them for misalignment. When filtering the vast majority of the content related to AI, we see significant decreases in misalignment rates. The opposite was also true - synthetic positive AI data led to self-fulfilling alignment.

    While post-training decreased the effect size, benign fine-tuning[1] degrades the effects of post-training, models revert toward their midtraining misalignment rates. Models pretrained on realistic or artificial upsampled negative AI discourse become more misaligned with benign fine-tuning, while models pretrained on only positive AI discourse become more aligned.

    This [...]



    ---

    Outline:

    (00:15) TL;DR

    (01:10) Abstract

    (02:52) Background and Motivation

    (04:38) Methodology

    (04:41) Misalignment Evaluations

    (06:39) Synthetic AI Discourse Generation

    (07:57) Data Filtering

    (08:27) Training Setup

    (09:06) Post-Training

    (09:37) Results

    (09:41) Base Models: AI Discourse Causally Affects Alignment

    (10:50) Post-Training: Effects Persist

    (12:14) Tampering: Pretraining Provides Alignment-In-Depth

    (14:10) Additional Results

    (15:23) Discussion

    (15:26) Pretraining as Creating Good Alignment Priors

    (16:09) Curation Outperforms Naive Filtering

    (17:07) Alignment Pretraining

    (17:28) Limitations

    (18:16) Next Steps and Call for Feedback

    (19:18) Acknowledgements

    The original text contained 1 footnote which was omitted from this narration.

    ---

    First published:
    December 20th, 2025

    Source:
    https://www.lesswrong.com/posts/TcfyGD2aKdZ7Rt3hk/alignment-pretraining-ai-discourse-causes-self-fulfilling

    ---



    Narrated by TYPE III AUDIO.

    ---

    Images from the article:

    Figure 1: An overview of our pretraining interventions. Training data discussing AI systems has a measurable effect on the alignment of LLMs prompted with “You are an AI assistant
    23 December 2025, 10:15 am
  • More Episodes? Get the App