Follow

A significant theme in my life is actively choosing not to Goodhart and then being punished for it

If an agent technically "cooperates" but computing the conditions for cooperation is computationally intractable then the agent isn't truly cooperating

Engaged in modal combat with institutions

Sign in to participate in the conversation
Mastodon

a Schelling point for those who seek one