How Machine Learning AI going to make your phone even smarter

We all know that our smartphones are as powerful as desktop computers from a few years ago, and even supercomputers from a few decades ago—and now they’re on the verge of teaching themselves to become even smarter and more competent. Today, image-recognition AI company Clarifai debuted new software that allows mobile apps to do machine learning—the actual training of models—right on smartphones, going back to at least the iPhone 5, with no cloud server farm required.
Understanding the significance of this development requires us to unpack some jargon. Tech companies frequently talk about apps that use machine learning—showing enough images or other data to an artificial intelligence program until it starts to discern important patterns, such as what a hamburger patty and bun look like. That’s how apps like Dog Breed Identifier or Food Calorie Counter, for instance, can identify what you point the phone’s camera at.
But these apps aren’t learning on the phone what, say, a Yorkie looks like. That process usually happens at a powerful server farm, often running on racks of graphics processing units (GPUs) such as NVidia’s Tesla P100 to develop a model of things like the identifying characteristics of a Yorkie. The apps on the phone apply that model to the particular dog in front of the camera, a process called inferring. It’s a processor-intensive operation, but not nearly as intensive as training a model.
Very few apps do the actual training on phones: Apple’s Photos app being a prime example for its limited ability to learn that, for instance, a particular face is yours or your friend’s. Clarifai introduced an app called Forevery in 2015 with similar capabilities. Apple’s upcoming iOS 11 has a component called Core ML that allows developers to add inference capability, but not training, to their apps. Google has taken a similar step with Android, announcing software called TensorFlow Lite for Android developers.
But with today’s release of software called Mobile SDK (software developer kit), Clarifai is letting developers build training for aspects of images and videos into any app. Clarifai’s CEO Matthew Zeiler explained the difference to me in an email: “You can train faces in . . . Apple Photos, but that’s it. With Clarifai’s Mobile SDK, we allow developers to train models on any set of concepts in their own apps (for example, a Lamborghini from a Ferrari, or a frappuccino from a latte, etc.) completely on device.”


Popular posts from this blog

What’s the difference between UI and UX?

Snap copied location sharing app Zenly to build Snap Map

IBM and SAP work together on artificial intelligence technology