The chief scientist leading Facebook's research into artificial intelligence doesn't believe the technology is a threat to human existence – unless we are very stupid.
"I don’t think at AI will become an existential threat to humanity," said Yann LeCun, the top professor who is leading Facebook's efforts in the space of AI and machine learning.
"I’m not saying that it’s impossible, but we would have to be very stupid to let that happen. Others have claimed that we would have to be very smart to prevent that from happening, but I don't think that's true," he said, in an in-depth question and answer session on Quora.
"If we are smart enough to build machines with super-human intelligence, chances are we will not be stupid enough to give them infinite power to destroy humanity," he continued.
"Also, there is a complete fallacy due to the fact that our only exposure to intelligence is through other humans. There are absolutely no reason that intelligent machines will even want to dominate the world and/or threaten humanity. The will to dominate is a very human one (and only for certain humans)."
Le Cun, who leads Facebook's AI Research (FAIR) added that behaviours specific to humans, such as violence or jealousy, would only be in intelligent machines if they are explicitly given to them. Even then, a second AI machine could be built with explicit instructions to destroy a dangerous AI.
"If both AIs have access to the same amount of computing resources, the second one will win, just like a tiger a shark or a virus can kill a human of superior intelligence."
"Microsoft is doing some good work, but they are losing quite a lot of people to Facebook and Google. They did some good work on deep learning in speech (and on handwriting recognition back in the early 2000s before the current craze). But their recent efforts seem to have been less ambitious than at FAIR or DeepMind," he said, when asked who was leading the way in research among the big tech firms.
"A number of top researchers from Google Brain have left for DeepMind, OpenAI, or FAIR," he added, while Apple's secretive nature meant they were "not a player in the AI research circuit".
"You simply cannot do leading-edge research in secret. If you can’t publish, it’s not research. At best, it’s technology development."
Asked whether the technology is experiencing a bubble, he said: "There is a lot of hype around AI and deep learning at the moment."
Read more: Can AI be creative? Google wants to find out
He went on to say: "That said, deep learning produces real results and is at the root of a real industry that makes money today. The promises of it in the near future are very exciting (even without the hype) in areas like self-driving cars, medical imaging, personalised medicine, content filtering/ranking, etc."
Facebook's AI research began two and a half years ago and the company makes much of its research publicly available. It recently launched a tool powered by AI which translates people's Facebook posts into different languages automatically.