Favorites
Report for Open Philanthropy examining what I see as the core argument for concern about existential risk from misaligned artificial intelligence.
Introduction and summary for a series of essays about how agents with different values should relate to one another, and about the ethics of seeking and sharing power.
There are oceans we have barely dipped a toe into. There are drums and symphonies we can barely hear. There are suns whose heat we can barely feel on our skin.
Infinities puncture the dream of a simple, bullet-biting utilitarianism. But they’re everyone’s problem.
How worried about AI risk will we be when we can see advanced machine intelligence up close? We should worry accordingly now.
My report examining the probability of a behavior often called “deceptive alignment.”
Making happy people is good. Just ask the golden rule.
I think that you can “control” events you have no causal interaction with, including events in the past, and that this is a wild and disorienting fact, with uncertain but possibly significant implications. This essay attempts to impart such disorientation.
If you kill something, look it in the eyes as you do.
On looking out of your own eyes.
You can’t keep any of it. The only thing to do is to give it away on purpose.
Nearby is the country they call life.
An intuition pump for a certain kind of “holy sh**” reaction to existential risk, and to the possible size and quality of the future at stake.
How can “non-attachment” be compatible with care? We need to distinguish between caring and clinging.