10 years ago a lot of AI Safety discussion turned on distilling humanity's meta-ethics into a machine-readable form. Today our most impressive AI's approximately reflect all the human content we could find for them, encoded in a semantically-meaningful way. We can convey intuitive preferences to the machines now. We can't guarantee that they'll actually optimize on those preferences, but the fact that the concepts are available seems under-discussed.
@jai From the early days, the plan was "figure out what human values are, and optimize for them". The problem has always been (1) how to encode that sentiment, and (2) how to build an AGI that reliably optimizes for *anything at all*, even eg maximising diamond.