Loud and clear: AI is improving Assistant conversations

Share
  • April 28, 2021

To get things done with the Google Assistant, it needs to understand you – it has to both recognize the words you’re saying, and also know what you mean. It should adapt to your way of talking, not require you to say exactly the right words in the right order.

Understanding spoken language is difficult because it’s so contextual, and varies so much from person to person. And names can bring up other language hiccups — for instance, some names that are spelled the same are pronounced differently. It’s this kind of complexity that makes perfectly understanding the way we speak so difficult. This is something we’re working on with Assistant, and we have a few new improvements to share.

Teach Google to recognize unique names 

Names matter, and it’s frustrating when you’re trying to send a text or make a call and Google Assistant mispronounces or simply doesn’t recognize a contact. We want Assistant to accurately recognize and pronounce people’s names as often as possible, especially those that are less common.

Starting over the next few days, you can teach Google Assistant to enunciate and recognize names of your contacts the way you pronounce them. Assistant will listen to your pronunciation and remember it, without keeping a recording of your voice. This means Assistant will be able to better understand you when you say those names, and also be able to pronounce them correctly. The feature will be available in English and we hope to expand to more languages soon.

A good conversation is all about context

Assistant’s timers are a popular tool, and plenty of us set more than one of them at the same time. Maybe you’ve got a 10-minute timer for dinner going at the same time as another to remind the kids to start their homework in 20 minutes. You might fumble and stop mid sentence to correct how long the timer should be set for, or maybe you don’t use the exact same phrase to cancel it as you did to create it. Like in any conversation, context matters and Assistant needs to be flexible enough to understand what you’re referring to when you ask for help.

To help with these kinds of conversational complexities, we fully rebuilt Assistant’s NLU models so it can now more accurately understand context while also improving its “reference resolution” — meaning it knows exactly what you’re trying to do with a command.  This upgrade uses machine learning technology powered by

These updates are now available for alarms and timers on Google smart speakers in English in the U.S. and expanding to phones and smart displays soon.

More natural conversations

We also applied BERT to further improve the quality of your conversations. Google Assistant uses your previous interactions and understands what’s currently being displayed on your smartphone or smart display to respond to any follow-up questions, letting you have a more natural, back-and-forth conversation.

Have a more natural, back and forth conversation with Google

If you’re having a conversation with your Assistant about Miami and you want more information, it will know that when you say “show me the nicest beaches” you mean beaches in Miami. Assistant can also understand questions that are referring to what you’re looking at on your smartphone or tablet screen, like [who built the first one] or queries that look incomplete like [when] or [from its construction]. 

There’s a lot of work to be done, and we look forward to continue advancing our conversational AI capabilities as we move toward more natural, fluid voice interactions that truly make everyday a little easier. 

Source : Loud and clear: AI is improving Assistant conversations