November 5, 2025
The Optimistic Assumption
Many AI safety discussions assume that Artificial Superintelligence (ASI) will be:
- Capable of solving problems humans can’t
- Able to reason about ethics and values
- Potentially omniscient (or close enough)
But …
November 4, 2025
Most AI risk discussions focus on x-risk: existential risk, scenarios where humanity goes extinct. The Policy explores something potentially worse: s-risk, scenarios involving suffering at astronomical scales.
The “s” stands for …
October 15, 2025
I asked an AI to brutally analyze my entire body of work—140+ repositories, 50+ papers, a decade and a half of research. The assignment: find the patterns I couldn’t see, the obsessions I didn’t know I had, the unifying thesis underlying …
October 13, 2025
Humanity has always fought against oblivion using stories, monuments, and lineage. But I no longer believe legacy will continue in that format. If something like Artificial Superintelligence endures beyond us, the mode of remembrance may shift from …
October 1, 2025
A speculative fiction novel exploring AI alignment, existential risk, and the fundamental tension between optimization and ethics. When a research team develops SIGMA, an advanced AI system designed to optimize human welfare, they must confront an …