Table of contents

Links are posted monthly in chronological order and include any articles, papers, or essays that I find pertinent to my current research and writing.



  • Dokkōdō
    • Penned by the famous Japanese swordsman Miyamoto Musashi in 1645, just a week before his death, Dokkōdō (which translates roughly to “The Way of Walking Alone” or “The Path of Aloneness”) offers 21 precepts or guiding principles for life.
  • Tiny Recursive Model (trm)
    • Instead of going big with tons of parameters, the designers built a two-layer model that recurses—meaning it iteratively refines a hidden “scratchpad” (latent space) plus a “current solution” over multiple steps. Think “draft + revise” loops rather than throwing everything in one long pass. They unroll this loop many times (up to 16) during training, supervise it “deeply” (i.e. give intermediate guidance), and allow full backpropagation through all those recursions (rather than using fixed-point approximations). Also, they ditch complex modular hierarchies (used by prior models) and instead do all the reasoning via that one compact recurrent core.
  • The Unbearable Lightness of Lightness
    • This essay explores lightness as both metaphor and existential feeling. Anna draws on personal experience, especially in relation to weightlifting or strength training (or absence thereof), to illustrate how being unable to “carry the iron” can feel like a loss of grounding. This absence of heaviness becomes a motif: when your body can’t bear weight, or you’re missing a part of your own routine, life feels insubstantial, untethered. She suggests that weight (in the literal or metaphorical sense) gives things substance and meaning. Without it, everything feels ethereal, faint—and yet paradoxically, that lightness can be unbearable. She echos the title of Kundera’s The Unbearable Lightness of Being and riffs on that tension between the meaningful weight of experience and the drift of what happens when weight is lost.
  • A trimodal protein language model enables advanced protein searches
    • ProTrek is built to understand three “languages” of proteins—sequence (the letters making up the protein), structure (its physical 3D shape), and natural language (textual descriptions like “enzyme that repairs dna”). It embeds all three modes into a shared latent space through clever contrastive learning. In practice, that means you could query a protein by giving any one of those modes and retrieve matches in any of the others.
  • Regression Language Models for Code
    • This paper pushes that idea in a slightly different direction: what if a language model could predict numeric properties of code—things like how much memory it uses, how fast it runs, or how accurate a neural net implemented in code will be?

  • The Self-Organization of Insight: Entropy and Power Laws in Problem Solving
    • The authors are trying to explain how people suddenly gain insight, which is the moment when something “clicks”, and a new structure or strategy emerges that lets one solve a problem in a qualitatively different way. They argue that traditional models in cognitive science (symbolic representations, fixed procedures, etc.) don’t do a great job capturing those abrupt, emergent changes. Instead, Stephen & Dixon suggest thinking of cognition (e.g. insight, problem solving) as a nonlinear dynamical system. This system is open, continuously interacting with the environment, where internal change (in thought, perception, action) can lead to self-organization of new cognitive structure. In this view, insight is not just “applying a known rule” but reorganizing how one is thinking, perceiving, acting.
  • Time100 ai 2025
    • According to time, these are the “innovators, leaders, and thinkers reshaping our world through groundbreaking advances in artificial intelligence.” I Think there are some very important people missing from this list, more related to research, that have shaped the landscape of AI.
  • Efficient nano-photonic antennas based on dark states in quantum emitter rings
    • Imagine you have a ring of quantum emitters—little things that can absorb light and later emit it (atoms, molecules, etc.). When many such emitters are arranged in a ring, interesting collective effects emerge. Some of the collective states are bright (they radiate strongly), and some are dark, meaning they radiate very little. They sort of trap excitation among the emitters instead of letting it leak out quickly. This paper studies how one can use a ring of emitters, together with an extra “absorber” placed at the center of the ring, to build a highly efficient nano-antenna that captures incoming light (or a photon) and absorbs it efficiently at the center, exploiting those dark states. The idea is that a dark state helps concentrate energy at the center (to be absorbed) without letting much be lost by radiation from the ring itself. They explore how many emitters in the ring works best, how the geometry and coupling matter, how detunings (mismatch in frequencies) affect things, etc. A particularly interesting result is that nine emitters in the ring (i.e. a nonagon) gives a “sweet spot” for absorption efficiency under many conditions. In many nanoscale and quantum optical technologies (e.g. light harvesting, single-photon detectors, quantum sensors), getting light in and absorbing or detecting it efficiently is hard. Losses by radiation, scattering, mismatch, etc. degrade performance. This ring + absorber design gives a way to reduce unwanted radiative losses (via the dark state) while still capturing energy effectively where you want it.
  • The entropic brain: a theory of conscious states informed by neuroimaging research with psychedelic drugs
    • This paper argues that conscious states vary along a spectrum of neural entropy. Normal waking consciousness is relatively constrained (less entropy), maintaining self, reality testing, etc. Psychedelics (and other primary consciousness states) raise entropy: more flexibility, more novel associations, less ego control, etc. Neuroimaging evidence from psilocybin (drops in dmn connectivity, drops in regular oscillatory power in some bands, increased metastability and variance, more diverse connectivity motifs) supports this. The theory has implications not just for how we understand consciousness, but for mental health and therapy.
  • Church Planting: When Venture Capital Finds Jesus
    • Evangelical church planting in the U.S. (especially non-denominational evangelical churches) contain many of the organizational, cultural, financial, and motivational patterns like startup / venture-capital / tech-entrepreneur ecosystems. Basically, church planting (starting new churches) has become a “venture-capital finds Jesus” world: the structure, incentives, growth-metrics, funder / planter relationships, risk / return thinking, charisma, etc., have strong analogues in startup culture. This post walks through what church planters are, how they are funded, what their goals are, how failures happen, what the costs are, and what the human side looks like (burnout, moral risk, etc.). Along the way there are comparisons to tech startups, VC dynamics, “hits-based” industries, etc.
  • Conversational Cultures: Combat vs Nurture (V2)
    • “Combat vs Nurture” is a proposed framing from Ruby that explains different conversational ecosystems with different trade-offs. Neither is “always right.” Depending on the group, topic, relational context, emotional stakes, one or the other (or a mix) is more appropriate.
  • Accelerating life sciences research
    • OpenAI has “successfully leveraged gpt4b micro to design novel and significantly enhanced variants of the Yamanaka factors, a set of proteins which led to a Nobel Prize for their role in generating induced pluripotent stem cells (iPSCs) and rejuvenating cells. They have also been used to develop therapeutics to combat blindness⁠(opens in a new window), reverse diabetes⁠, treat infertility⁠, and address organ shortages⁠.”
  • Collective alignment: public input on our Model Spec
    • Oai surveyed over 1,000 people worldwide on how their models should behave and compared user’s views to the current Model Spec. They found that they largely agree with the Spec, and changes were adopted from the disagreements.”
  • The Ultra-Scale Playbook: Training llms on gpu Clusters
    • If you’re someone who’s built a small model, or done training on a single gpu or a small multi-gpu rig, and wondered how people train massive models with 70b+ or 400b+ parameters, this playbook is like a guided map through that terrain. It lays out the tools, tricks, trade-offs, and gives you real data and code you can use. It doesn’t promise an exact recipe that works everywhere, but it gives you the intuition to design your own recipe, measure what matters in your setting, and avoid some of the big surprises: memory bottlenecks, wasted gpu time, communication lag, etc. If you’re serious about scaling up llm training, it’s one of the better resources out there right now.
  • The Rationality Trap
    • This is a sharp, introspective, and somewhat somber exploration of how the rationality movement, founded on the aspiration to think better and save humanity, drifted into dangerous territory. It doesn’t paint everything in the rationalist / AI safety world as bad, but Weiss-Blatt shows that certain mythologies, internal methods, stakes framing, and funding structures may have created fertile ground for cultish dynamics.
  • This is the World’s Biggest Animal Migration
    • Scientists have used aerial surveys and tracking to document a migration of millions of antelope (white-eared kob, tiang, Mongalla gazelle, bohor reedbuck) moving across huge distances in search of water and grazing. This is now being called the Great Nile Migration—possibly the largest land mammal migration on Earth, both in sheer numbers and spatial scale.
  • Contemplative Artificial Intelligence
    • “As artificial intelligence (AI) improves, traditional alignment strategies may falter in the face of unpredictable self-improvement, hidden subgoals, and the sheer complexity of intelligent systems. Inspired by contemplative wisdom traditions, [the paper] shows how four axiomatic principles can instil a resilient Wise World Model in AI systems. For future systems, active inference may offer the self-organizing and dynamic coupling capabilities needed to enact Contemplative AI in embodied agents.”
  • The Illusion of Diminishing Returns: Measuring Long Horizon Execution in llms
    • “Failures of llms when simple tasks are made longer arise from mistakes in execution, rather than an inability to reason. We propose isolating execution capability, by explicitly providing the knowledge and plan needed to solve a long-horizon task. We find that larger models can correctly execute significantly more turns even when small models have 100% single-turn accuracy.”
  • RL Research

  • Silk Road Timeline
  • Satiety Graphed & The Horsemen of Obesity
    • This is a criticism of the common appeal to satiety without a clear definition, arguing that relying on an undefined concept is problematic, even if, by chance, a diet works. Satiety, not to be confused with physical fullness, can indicate metabolic and dietary health, and knowing it when you see it can work as a pragmatic definition.
  • Gpt-oss release
    • Ollama: OpenAI gpt-oss
    • Welcome gpt oss, the new open-source model family from OpenAI!
      • Gpt oss is a hugely anticipated open-weights release by OpenAI, designed for powerful reasoning, agentic tasks, and versatile developer use cases. It comprises two models: a big one with 117b parameters (gpt-oss-120b), and a smaller one with 21b parameters (gpt-oss-20b). Both are mixture-of-experts (MoEs) and use a 4-bit quantization scheme (mxfp4), enabling fast inference (thanks to fewer active parameters, see details below) while keeping resource usage low. The large model fits on a single H100 gpu, while the small one runs within 16gb of memory and is perfect for consumer hardware and on-device applications.
    • Open models by OpenAI
      • Advanced open-weight reasoning models to customize for any use case and run anywhere.
  • Lagerstroemia
    • commonly known as crape myrtle (also spelled crepe myrtle or crêpe myrtle), is a genus of deciduous and evergreen trees and shrubs native to the Indian subcontinent, southeast Asia, northern Australia, and other parts of Oceania, cultivated in warmer climates around the world.
    • I’ve been taking care of a crepe myrtle on my property for several years and quite enjoy seeing it mature from a sapling. A year after moving into my current house, the tree didn’t recover after a harsh winter, so I was forced to cut the tree down, but I left the root system. In a few months, a few saplings began to grow, and after a lot of care and attention, it now has 5 main trunks measuring 6–7 inches in circumference and is over 15ft tall.
  • Oral treatment of erectile dysfunction with apomorphine SL
    • Apomorphine SL (Ixense, Uprima) is a new oral medication shown to be effective in the treatment of erectile dysfunction. This compound is a dopaminergic agonist with affinity for dopamine receptor sites—mostly D(2)—within the brain known to be involved in sexual function. Apomorphine induces selective activation in the nucleus paraventricularis leading to erectogenic signals.
  • Psilocybin treatment extends cellular lifespan and improves survival of aged mice
    • Psilocybin, the naturally occurring psychedelic compound produced by hallucinogenic mushrooms, has received attention due to considerable clinical evidence for its therapeutic potential to treat various psychiatric and neurodegenerative indications. However, the underlying molecular mechanisms remain enigmatic, and few studies have explored its systemic impacts. We provide the first experimental evidence that psilocin (the active metabolite of psilocybin) treatment extends cellular lifespan and psilocybin treatment promotes increased longevity in aged mice, suggesting that psilocybin may be a potent geroprotective agent.

  • Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task

    • The study investigates the neural underpinnings of creative writing across different participant groups: “Brain-only” (unassisted), “llm” (Large Language Model-assisted), and “Search Engine” (search engine-assisted). A primary focus is on Delta band connectivity, which consistently shows the most significant disparities between groups. The Brain-only group demonstrates significantly higher and more widespread Delta band activity, indicating greater engagement of deep, slow integrative brain processes, multisensory integration, and internally-driven thought during unassisted writing. Conversely, llm and Search Engine groups exhibit more externally anchored or intermittently guided cognitive engagement, with notably weaker delta interactions.
    • N-gram analysis complements these neurological findings, revealing distinct linguistic patterns across groups for different topics. The “Brain-only” group tends to use n-grams reflecting internal thought processes and prosocial framing, while llm and Search Engine groups show patterns indicative of external sourcing or more direct task-oriented language. The study also tracks changes in brain connectivity patterns across sessions, noting a general trend of increasing connectivity in later sessions for both Brain-only and llm groups, suggesting adaptation to the task
  • Ultrafast coherent dynamics of microring modulators

    • Next-generation computing clusters require ultra-high-bandwidth optical interconnects to support large-scale artificial-intelligence applications. These electronic–photonic co-integrated systems necessitate densely integrated high-speed electro-optical converters. In this context, microring modulators (mrms) emerge as a promising solution, prized for their exceptional compactness and energy efficiency.
  • Hitchhiker’s Guide to rag: From Tiny Files to Tolstoy with OpenAI’s api and LangChain

    • Scaling a simple rag pipeline from simple notes to full books
  • goose

    • an open source, extensible AI agent that goes beyond code suggestions—install, execute, edit, and test with any llm
  • If--

    • Kipling’s poem “If—” is essentially a father’s advice to his son (and, by extension, to any young person) on how to live a balanced, honorable life. Written in 1895 and later included in his 1910 collection Rewards and Fairies, it lays out a series of “if” scenarios—tests of character—and promises the ultimate reward (“you’ll be a Man, my son!”) if you can meet them all.
  • bitchat-android

    • bluetooth mesh chat, irc vibes
  • Potemkin Understanding in Large Language Models

    • Success on benchmarks only demonstrates potemkin understanding: the illusion of understanding driven by answers irreconcilable with how any human would interpret a concept.
  • The Best Tacit Knowledge Videos on Every Subject

Much of my time this month has been spent on researching tools that support the development of this site and implementing visual and basic quality-of-life features to establish a solid foundation for the future.

KR

New Links & Articles: newsletter & rssRSS icon