AI Won’t Replace Us- The late Stephen Hawkins worried that AI could end mankind. It seemed reasonable. Elon Musk warned machines that learned to operate without a human telling them what to do could “destroy humanity as a matter of course without even thinking about it” if it “[had] a goal and humanity just happens [to be] in the way.”.
But reality has proven that while AI can beat humans at games, it still fails at common tasks an infant can do, such as holding an object. In fact, to solve this problem, researchers from OpenAI used 6144 CPUs and 8 GPUs to collect about one hundred years of experience and trained the AI for 50 hours. As a result, the robotic hand can handle unknown objects — as long as they are “within reason.”
The fundamental gap
As Antonio Bicchi, a professor of robotics at the Istituto Italiano di Tecnologia said, the research had a number of limitations, such as the hand is always facing up so that the objects always fall in the palm.
We can’t tell for certain if another 100 years of training data would make the AI even better, or if it needs a new set of training data. What we can say is that humans are exceptionally good at incremental learning. Once a human learns to play with a ball, they can master any ball game quite easily. Or when we learn a foreign language, learning other languages becomes easier every time.
But an AI must learn everything from scratch. AI can’t use “other AIs.” It always starts from zero. AI cannot be “combined” with other AIs to do more complex tasks — at least not yet. So, while AI masters skills at a superhuman level, it only masters one task.
The missing piece
Recent developments in AI were boosted with the invention of deep learning algorithms and improved computer powers, which seemed to mimic how the brain operates by simulating perceptrons. But the brain is much more than that.
We don’t know how the real brain works and, according to Sam Rodriques, we’ll never know until we drill holes in the human skull and plant probes to study how it really works “behind the scenes” (or bones). But what we do know is that the brain is much less rational than we used to think. Studies prove that we decide first, then we try to find reasons for why we decided the way we did.
In fact, people with brain damage who were incapable of developing emotions could perfectly describe what they should be doing in logical terms, yet they found it very difficult to make even simple decisions, such as what to eat. Our choices are arguably always based on emotion.
Yet, there is no AI system that works like this. AI can’t reason, which has led to hidden AI bias in many projects. Of course, there is work in progress to solve these cases, but AI was primarily designed as a black box since we did not know how to code it in the first place.
In addition, just throwing more power in and building bigger machines is not the way if the machine takes the wrong path, and it can be expensive to know where it is really going since we can’t know what it has learned.