A significant theme in my life is actively choosing not to Goodhart and then being punished for it
If an agent technically "cooperates" but computing the conditions for cooperation is computationally intractable then the agent isn't truly cooperating
Engaged in modal combat with institutions
:surprised_pikachu:
a Schelling point for those who seek one
If an agent technically "cooperates" but computing the conditions for cooperation is computationally intractable then the agent isn't truly cooperating
Engaged in modal combat with institutions