> And just—back in the 'aughts, Robin Hanson had this really great blog called _Overcoming Bias_. (You probably haven't heard of it, I said.) I wanted that _vibe_ back, of Robin Hanson's blog in 2008—the will to _just get the right answer_, without all this galaxy-brained hand-wringing about who the right answer might hurt.
Man how can MTSW rock so much
I'm sorry I can't tell you the name of God. I can't tell you the name of God because he is veiled and almightly over in the heavens. he is veiled and almightly over in the heavens because you have exiled him there. You have exiled him there because you were afraid of your good twin. You were afraid by your good twin because your heart is burning off
more towards zero than one (so <<5%). Obviously the value-laden part wouldn't be solved.
I think the philosophy/math/cs system would be just as capable at capabilities work as at alignment work.
But I now remember an old idea of making STEMGPT, trained (in the weak case) only on STEM textbooks, arXiv, (in the strong case only on) hadron collider data, protein structures, meteorological and geological data &c. Hard to have info about humans leak over though.
How much of strawberry alignment is value-laden? 5%? 95%? probably further along some logarithmic scale, if I had to bet
I operate by Crocker's rules[1].