Google makes its AI assistant more accessible with 'Look and Talk'

Instead of saying 'Hey, Google' you'll have to make eye contact with a Nest Hub camera.

Sponsored Links

The Google Nest Hub Max, a smart home display, being exhibited at the Google Smart Home in the Android Avenue during the Mobile World Congress (MWC) the biggest trade show of the sector focused on mobile devices, 5G, IOT, AI and big data, celebrated in Barcelona, on March 3, 2022 in Barcelona, Spain.
 (Photo by Joan Cros/NurPhoto via Getty Images)
NurPhoto via Getty Images

Google Assistant is already pretty handy, filling in your payment info on take out orders, helping get the kids to school on time, controlling your stereo systems' volume and your home's smart light schedules. At its I/O 2022 keynote today, company executives showed off some of the new features arriving soon for the AI.

The first of these is "Look and Talk." Instead of having to repeatedly start your requests to Assistant with "Hey Google," this new feature relies on computer vision and voice matching to constantly pay attention to the user. As Sissie Hsiao, Google's VP of Assistant, explained on stage, all the user has to do is look at their Nest Hub Max and state their request. Google is also developing a series of quick commands that users will be able to shout out without having to gaze longingly at their tablet screen or say "Hey Google" first — things like "turn on the lights" and "set a 10-minute alarm."

asdf
Alphabet

All of the data captured in that interaction — specifically the user's face and voice prints, used to verify the user — are processed locally on the Hub itself, Hsiao continued, and not shared with Google "or anyone else." What's more, you'll have to specifically opt into the service before you can use it.

According to Hsiao, the backend of this process relies on a half-dozen machine learning models and 100 camera and mic inputs — i.e., proximity, head orientation and gaze direction — to ensure that the machine knows when you're talking to it versus talking in front of it. The company also claims that it worked diligently to make sure that this system works for people across the full spectrum of human skin tones. 

Looking ahead, Google plans to continue refining its NLP models to further enhance the responsiveness and fidelity of Assistant's responses by "building new, more powerful speech and language models that can understand the nuances of human speech," Hsiao said. "Assistant will be able to better understand the imperfections of human speech without getting tripped up — including the pauses, 'umms' and interruptions — making your interactions feel much closer to a natural conversation."

Follow all of the news from Google I/O 2022 right here!

All products recommended by Engadget are selected by our editorial team, independent of our parent company. Some of our stories include affiliate links. If you buy something through one of these links, we may earn an affiliate commission.
Popular on Engadget