Is AI a game for the big dogs?
Artificial Intelligence is becoming a game for the big dogs. To succeed in genuine artificial intelligence efforts, beyond trivial prediction, you need three things.
Sign-up to Exponential View, my weekly newsletter.
The first is people. AI is hard. AI is emerging, so you need to be able to recruit and attract a sufficient core of people and create the culture that allows them to explore. You also need mechanisms to exploit their progress commercially.
The second is training environments, which include training data but also systems which let you train and test your systems. The web now offers tons of labelled training data for images and text, with all of its attendant biases. There are data sets for self-driving cars available too.
This week researchers thinking about reinforcement learning or AI systems that can plan ahead received a double fillip. OpenAi, the non-profit backed by Elon Musk, released Universe a wide-ranging environment for researchers to develop generalised game playing systems. DeepMind released a similar environment. Both of these will lower the cost of research and experimentation.
The third is plank is processing. Moore’s Law has been our friend for 40 years, in recent years architectural improvements have helped drive down the cost of computing, particular the matrix operations so crucial in neural net algorithms. However, the computational complexity needed to afford modern breakthroughs is increasing substantially.
Training a simple classifier might only take a few GPU hours. But newer systems that can do more just require so much more computational horsepower. And those horses will come at a cost.
Discussing this with one AI researcher, we noted that it isn’t uncommon for the computational cost to train a modern set of algorithms might run into the tens, if not hundreds of thousands of dollars.
Example: Deepmind’s AlphaGo required 1202 CPUs and 176 GPUs to play Lee Sedol. Making some very crude estimates of cost… if you were to rent that level of processing from Amazon Web Services at 10 cents per hour, a very rough mid-point for a single reasonable processor – then this rig would cost nearly $200 per hour to run. And typically the ‘inferencing’ step is computationally cheaper than training AI systems.
Yes – the estimates are only directionally correct, we should work on more accurate ones.
But the point is that each of those ingredients people, data and compute are hard. But it could be increasingly the case that compute is the hardest & most expensive. As Andrew Ng, lead AI researcher of Baidu, recently said “There are multiple experiments I’d love to run if only we had a 10-x increase in performance.“ A 10-x performance improvement will take 3-5 years to drop to today’s already high prices.
So a handful of superpowers emerging in real frontier artificial intelligence, among them Google, Amazon, Apple, Baidu, Facebook, Microsoft and, increasingly, Uber.
- Apple may be struggling to get adoption of its deep learning toolkits. It took the unusual step of starting to publicise some of its research, and announced a willingness to collaborate. (Some juicy details on Apple’s technical progress.)
- Baidu has been making great strides under Andrew Ng’s leadership. Here Ng argues that an AI winter is not coming.
- Uber is proving rather visionary. By some estimates about a thousand of its 3,000 engineers are involved in machine learning. This week it acquired Geometric Intelligence, an AI startup that focusses on solving the problem of learning without needing scads of data.
What role for smaller companies? I'll return to that question in a future post.
Enjoy this? Sign-up to Exponential View, my weekly newsletter.
Digital Marketing & Communications | Strategy | Social Media | Advertising | Project Management
7yYes! Great article. To build artificial intelligence, you first need REAL intelligence! This is the true brain power behind rising software trends like analytics: https://www.gensuite.com/products-and-services/insight-analytics/
--
7ynice
Interesting, 1202 CPUs and 176 GPUs to beat a GO master, but then he can then go and dress himself which the AI could not do (without additional processors and training).
System Administrator at MicroStrategy
7ygoo.gl/6FxAw5
This "big dog" analogy makes sense for deploying some resources but not others. Certainly the expansion capital resource is a given. But how about the people (the key ingredient)that make it happen? I believe that many of the best suited to drive AI will be totally turned off by big company brand delivery hiring practices. Those companies that understand this and have the confidence to act on real qualifications will be the winners.