Deep Learning AI Becomes Portable.
Artificial intelligence programs are getting smarter and already they are meeting or exceeding what human experts can do in areas like chess, games of skill and mathematics computations. AI using deep learning protocols is on a trajectory to exceed what the human brain can do, threatening millions of jobs long before we have a chance to prepare.
At least for now these AI programs are constrained inside a rather large supercomputer. This however may also be changing quickly.
Technology developed at the University of Waterloo is paving the way for artificial intelligence (AI) to break free of the internet and cloud computing.
New deep learning AI software produced with that technology is compact enough to fit on mobile computer chips for use in everything from smartphones to industrial robots.That would allow devices to operate independent of the internet while using AI that performs almost as well as tethered neural networks.
Alexander Wong, a systems design engineering professor and Waterloo and co-creator of the technology, said:
We feel this has enormous potential. This could be an enabler in many fields where people are struggling to get deep-learning AI in an operational form.
The use of stand-alone deep learning AI could lead to much lower data processing and transmission costs, greater privacy and use in areas where existing technology is impractical due to expense or other factors.
Deep-learning AI, which mimics the human brain by processing data through layers and layers of artificial neurons, typically requires considerable computational power, memory and energy to function.
Researchers took a page from evolutionary forces in nature to make that AI far more efficient by placing it in a virtual environment, then progressively and repeatedly depriving it of resources.
The deep learning AI responds by adapting and changing itself to keep functioning each time computational power and memory are taken away.
“These networks evolve themselves through generations and make themselves smaller to be able to survive in these environments,” said Mohammad Javad Shafiee, a systems design engineering research professor at Waterloo and the technology’s co-creator.
In work recently presented during the International Conference on Computer Vision in Venice, Italy, the researchers achieved a 200-fold reduction in the size of deep-learning AI software used for a particular object recognition task.
When put on a chip and embedded in a smartphone, such compact AI could run its speech-activated virtual assistant and other intelligent features, greatly reducing data usage and operating without internet service.
Other potential applications range from use in low-cost drones and smart grids, to surveillance cameras and manufacturing plants, where there are significant issues around streaming sensitive or proprietary data to the cloud.
Wong and Shafiee, who have co-founded a company called DarwinAI to commercialize their efficient AI software, were “amazed” at the results when they first attempted their approach to evolving deep-learning AI about three years ago.
“We are researchers, so we explore many different things,” said Shafiee. “And if it works, we keep going and push harder.”