This is potentially related to my other question, answers to one could theoretically affect the other.
Hard AI, that is to say thinking AI that has the ability to reason, plan, and learn the way humans do, is a very common trope in science fiction. Discussion of how reliable these AI will be have been one of the biggest concepts.
We've heard enough of the "AI turns against their masters stories" that humans will consider the concept when constructing hard AI. I'm looking to understand what steps we would take in the development and use of AI to ensure such a scenario doesn't happen, and ensure the AI is worth constructing for humans.
One option is we use no approaches and build AI to be identical to humans. Another commonly discussed alternative is Asimov's 3 laws of robotics. Other ideas like constraining bolts exist, some way to enforce a robot/AI to behave in a given manner even if against it's 'will', or just crippling it's growth to prevent it from ever reaching the degree of true sapient required to have a 'will' of it's own. While it doesn't seem to be seen as much in sci-fi, perhaps because it offers less interesting story telling options, another idea is to program AI to have an innate desire to help that predisposes them working with and supporting humans (imagine if AI simply feel happy about helping and sad about hurting to such a degree that they want to help people always because it makes them feel 'good'; an artificial conscious to go with their artificial intelligence that happens to be ramped up to 10 to ensure altruism).
There are no doubt other approaches out there I haven't mentioned as well that ensure Hard AI cooperation, and of course I'm interested in them as well.
My question is, assuming we eventually create Hard AI, what solution, or solutions, will we utilize to ensure cooperation and support of the AI?
The only other question that asks this directly assumes a very specific situation, including that an antagonistic AI already existed and were retroactively trying to control it. Similar questions that touch on control AI likewise seem to have a same problem to me, they don't feel like their focused on realistic development of AI beyond presuming it exists. That's why This question is marked hard-science, I want to focus exclusively on the most likely avenues in a real world scenario, how our AI will be developed originally and how the growth of AI cooperation techniques may occur along side development of AI.
I'm looking for answers that focus on realistic AI development when considering what approaches for ensuring cooperation are practical. Ideally answers will not just consider the science, but also how human nature may affect development, including human prejudice or short-sightedness, or empathizing with sapient AI and wish to avoid their suffering, or basic economics & cost/reward analysis which could affect approaches utilized in AI development.
I wouldn't be surprised if many answers at least touch on what approach we take to create Hard AI briefly in order to address why a given mechanisms make sense. Answers should touch on why such mechanisms work, and/or the limits to how far they can go. Keep in mind reality checks mean ensuring any purposed AI is desirable to build for humans to begin with; AI either too crippled to be useful or too likely to betray or abandon humans after being constructed would presumably not be worth the expense of building unless we didn't realize the issues until after we had already created them.