phone Google Assistant, in the nature and manner of
Google Now, can search the Internet, schedule events and alarms, adjust hardware settings on the user's device, and show information from the user's Google account. Unlike Google Now, however, the Assistant can engage in a two-way conversation, using Google's
natural language processing algorithm. Search results are presented in a card format that users can tap to open the page. In February 2017, Google announced that users of Google Home would be able to shop entirely by voice for products through its
Google Express shopping service, with products available from
Whole Foods Market,
Costco,
Walgreens,
PetSmart, and
Bed Bath & Beyond at launch, and other retailers added in the following months as new partnerships were formed. Google Assistant can maintain a
shopping list; this was previously done within the notetaking service
GKeep, but the feature was moved to
Google Express and the Google Home app in April 2017, resulting in a severe loss of functionality. In May 2017, Google announced that the Assistant would support a keyboard for typed input and visual responses, support identifying objects and gather visual information through the device's camera, and support purchasing products and sending money. Through the use of the keyboard, users can see a history of queries made to the Google Assistant, and edit or delete previous inputs. The Assistant warns against deleting, however, due to its use of previous inputs to generate better answers in the future. In November 2017, it became possible to identify songs currently playing by asking the Assistant. The Google Assistant allows users to activate and modify vocal
shortcut commands in order to perform actions on their device (both Android and iPad/iPhone) or configure it as a hub for
home automation. This feature of the
speech recognition is available in English, among other languages. In July 2018, the Google Home version of Assistant gained support for multiple actions triggered by a single vocal shortcut command. At the annual
I/O developers conference on May 8, 2018, Google's
SEO announced the addition of six new voice options for the Google Assistant, one of which being
John Legend's. This was made possible by WaveNet, a voice synthesizer developed by DeepMind, which significantly reduced the amount of audio samples that a voice actor was required to produce for creating a voice model. However, John Legend's Google Assistant cameo voice was discontinued on March 23, 2020. In August 2018, Google added bilingual capabilities to the Google Assistant for existing supported languages on devices. Recent reports say that it may support multilingual support by setting a third default language on Android Phone. Speech-to-Text can recognize commas, question marks, and periods in transcription requests. In April 2019, the most popular
audio games in the Assistant, Crystal Ball, and Lucky Trivia, have had the biggest voice changes in the application's history. The voice in the assistant has been able to add expression to the games. For instance, in the Crystal Ball game, the voice would speak slowly and softly during the intro and before the answer is revealed to make the game more exciting, and in the Lucky Trivia game, the voice would become excitable like a game show host. In the British accent voice of Crystal Ball, the voice would say the word 'probably' in a downward slide like she's not too sure. The games used the text-to-speech voice which makes the voice more robotic. In May 2019 however, it turned out to be a bug in the speech API that caused the games to lose the studio-quality voices. These audio games were fixed in May 2019.
Interpreter Mode On December 12, 2019, Google debuted an interpreter mode in Google Assistant smartphone apps for Android and iOS. It provides translation of conversations in real-time and was previously only available on Google Home smart speakers and displays. Google Assistant won the
2020 Webby Award for Best User Experience in the category: Apps, Mobile & Voice. On March 5, 2020, Google introduced a feature on Google Assistant that read webpages aloud in 42 languages. On October 15, 2020, Google announced a new 'hum to search' function to find a song by simply humming, whistling, or singing the song.
Google Duplex In May 2018, Google revealed Duplex, an extension of the Google Assistant that allows it to carry out natural conversations by mimicking human voice, in a manner not dissimilar to
robocalling. The assistant can autonomously complete tasks such as calling a hair salon to book an appointment, scheduling a restaurant reservation, or calling businesses to verify holiday store hours. While Duplex can complete most of its tasks fully autonomously, it is able to recognize situations that it is unable to complete and can signal a human operator to finish the task. Duplex was created to speak in a more natural voice and language by incorporating
speech disfluencies such as
filler words like "hmm" and "uh" and using common phrases such as "mhm" and "gotcha", along with more human-like
intonation and response latency. Duplex is currently in development and had a limited release in late 2018 for Google Pixel users. During the limited release, Pixel phone users in Atlanta, New York, Phoenix, and San Francisco were only able to use Duplex to make restaurant reservations. As of October 2020, Google has expanded Duplex to businesses in eight countries.
Criticism After the announcement, concerns were made over the ethical and societal questions that
artificial intelligence technology such as Duplex raises. For instance, human operators may not notice that they are speaking with a digital robot when conversing with Duplex, which some critics view as unethical or deceitful. Concerns over privacy were also identified, as conversations with Duplex are recorded in order for the virtual assistant to analyze and respond. Privacy advocates have also raised concerns around how the millions of vocal samples gathered from consumers are fed back into the algorithms of virtual assistants, making these forms of AI smarter with each use. Though these features individualize the user experience, critics are unsure about the long term implications of giving "the company unprecedented access to human patterns and preferences that are crucial to the next phase of artificial intelligence". While transparency was referred to as a key part to the experience when the technology was revealed, Google later further clarified in a statement saying, "We are designing this feature with disclosure built-in, and we'll make sure the system is appropriately identified." == Reception ==