Sunday, September 25, 2016

What's in that photo? Google open-sources caption tool in TensorFlow that can tell you

#Google has open-sourced a model for its machine-learning system, called #ShowandTell, which can view an image and generate accurate and original captions.

The model it's released is faster to train and better at captioning images than the versions that previously helped it secure a tied first place with #MicrosoftResearch in #Microsoft 's COCO 2015 image-captioning contest.

The image-captioning system is available for use with #TensorFlow, Google's open machine-learning framework, and boasts a 93.9 percent accuracy rate on the ImageNet classification task, inching up from previous iterations.

The code includes an improved vision model, allowing the image-captioning system to recognize different objects in images and hence generate better descriptions.

An improved image model meanwhile aids the captioning system's powers of description, so that it not only identifies a dog, grass and frisbee in an image, but describes the color of grass and more contextual detail.

The improvements, detailed in a new paper, apply recent advances in computer vision and machine translation to image-captioning challenges. Google researchers see potential for it as an accessibility tool for visually-impaired people when viewing images on the web.
http://www.zdnet.com/article/whats-in-that-photo-google-open-sources-caption-tool-in-tensorflow-that-can-tell-you/

No comments:

Post a Comment