Discussion about this post

User's avatar
Noah Birnbaum's avatar

Good piece!

I think this is actually a subset of the alignment problem as opposed to a different problem entirely. Part of one’s values is their meta-values (ie the things they would want themselves to want). It seems like you’re arguing that AI can change what we want (as have the social media algorithms, etc). However, what we want to want is merely just a preference that AI is then being misaligned to.

I do like the framing you give a lot.

One issue that I have with this piece, though, is that it frames this as a larger problem than alignment (and the associated risks) itself without much justification - while I agree that this type of thing is a problem (and can be somewhat distinguished from alignment as it’s normally stated), I don’t see why it would be a bigger deal than the potential risks associated with super capable misaligned AI systems.

Winston Margaritis's avatar

Excellent work. I especially enjoyed this paragraph:

"The Christian (and the Buddhist) is supposed to mortify the flesh, to put to death the desires that enslave him and thus pull him away from God. Christianity requires repentance, and part of Christianity’s problem in the West is that people don’t recognize how impoverished is modern life. “We are far too easily pleased.” Sin isn’t always or even usually dramatic. Rather, sin is usually soporific. It lulls you asleep. ‘Why do I need to repent? I’m a pretty good person with a pretty good life’—say the damned."

2 more comments...

No posts

Ready for more?