You are thinking in narrative terms about AI. Crazy <-> Rational is a far more likely issue with early AI than Good <-> Evil. The problem is you can't optimize for general intelligence only solutions to given problems. Let's say you have two twins one of which is smarter than the other. If you ask about future predictions then how do you know the correct answer before hand.
In other words the 99% of paperclip optimizers are going to start making virtual paperclips not paving over the universe with actual paperclips. Hacking your reward function is easer than solving hard problems.
In other words the 99% of paperclip optimizers are going to start making virtual paperclips not paving over the universe with actual paperclips. Hacking your reward function is easer than solving hard problems.
PS: Don't do drugs :0