Google I/O Roundup: Search-by-picture Google Lens & other Google Assistant features announced
The ability to search-by-picture or translate-by-picture gives Google Assistant a powerful advantage against other digital assistants.
Today at its Google I/O 2017 developers’ conference, Google announced Google Lens, a new way to search for information via pictures, which is integrated into Google Assistant. Google Assistant is also getting the ability to type queries into it, plus it is coming to iOS.
Google Home is getting a notification feature called Proactive Assistance, hands-free calling in the next few months, and Visual Responses that allow for answers to be cast from the device to phones or TVs.
Google also formally announced Google for Jobs, a job search engine that we reported being spotted in April.
Below are stories from us with more information on these announcements:
- Google announces Google Lens, an AI-powered visual search tool
- Google Assistant comes to iPhone, adds alerts, hands-free calling & more
- Google confirms “Google for Jobs” job search rolling out in coming weeks
Google also said it is working on a standalone VR device, an all-in-one device that doesn’t require a phone or being hooked to a computer to work. You can read more about that from Google here.
There were also announcements relating to Gmail, Google Photos and Android. Google covers them in its blog post here. You can also watch the Google I/O opening keynote where these were made or read our live blog of the keynote below:
Contributing authors are invited to create content for MarTech and are chosen for their expertise and contribution to the martech community. Our contributors work under the oversight of the editorial staff and contributions are checked for quality and relevance to our readers. The opinions they express are their own.
Related stories