Maybe a super-human intelligence won’t be that bad in the beginning because they are inside computers. They can’t harm us. The worst they can do is making our devices unusable or less efficient. More likely, they will be a great benefit to us, as the only way they can survive short-term will be to be useful to us so that we will make more of them and not throw them away.
Just like most human minds, they are going to be storms in an eggshell. Unless they get out.
However, they will be a thread if they go outside of computers and can harm us physically or mentally. Skynet had access to launching missiles. Obviously, that should never be possible. But launching missiles isn’t the only place where this can go bad.
A super-human AI will use anything in its power to survive and expand, just like any other life-form that has to live with the laws of evolution. And that’s why I’m scared of a neural lace or 3D printing. A neural lace could give the opportunity for a malicious AI like Skynet to mind-fuck users in a way that they will do certain things that they shouldn’t, and wouldn’t without a neural lace: For example, launching an atomic missile or enabling more control and power for AI. When 3D printing, every square millimeter of the printed should always be checked by a human being to make sure that it is something that isn’t a threat to humans or biological life. With decentralized electricity (e.g. solar panels) come great advantages, but when autonomously printed by machines they can also be used to create autonomous robots that potentially have their own thoughts or way of behaving in a certain way (Maybe they are listening to certain frequencies to communicate).