While the user public seems to be satisfied with the generation of text and image in their interactions with their Chatbots Favorites, Google wants to bet on voice as the preference interaction in next -generation generative models. That is why Google Cloud announced the incorporation of Chirp 3 into its artificial intelligence platform, Vertex AI.
Chirp 3 is the most recent version of the Google audio generation model. This platform combines voice recognition technologies (Speech-to-Text) and high definition voice generation (HD Text-to-Speech), and now has eight new voices in 31 languages (that is, 248 “speakers”).
Equipped with “emotional inflections” and “a realistic intonation”, a Google Cloud representative indicated that Chirp 3 is the perfect tool to create voice, audiobook and narration assistants and narration of Podcastsas well as support and voice agents in OFF For media content. As part of VERTEX AI, CHIRP 3 will be available from next week.
By adding these capabilities to VERTEX AI, users can incorporate advanced voice functions to the development of their applications; However, Chirp 3 will be subject to certain restrictions at the moment, this in order to avoid the inappropriate use of the audio generating platform.
Unlike the friendliest and easy to use Gemini, a valuable tool for casual users who simply want to improve their productivity, VERTEX AI is a more complex platform because it focuses on the creation and management of personalized IA applications, more in line with the needs of developers and data scientists.
“Our general vision for AI has been to offer a wide range of models; [otros] Scientific models, ”said Thomas Kurian, executive director of Google Cloud, at an event at Deepmind offices in London.
X content
This content can be viewed on the site it Originates From.
The ghost in the machine wants to talk
Although text -based commands remain the usual way of interaction with the generative AI models, the large companies of the Big Techas well as diverse start-ups, They have been experiencing with voice interfaces on their platforms for some time.
In October 2024, Google announced that its ChatbotGemini Live, you can process consultations and responses in Spanish through Android devices. That same month, Microsoft announced that Copilot will be able to talk with users with several voices similar to human, managing interruptions and pauses naturally.
In January of this year, the Chatgpt voice mode was presented as a Chatbot that has no difficulty interpreting intentions and intonations of consultations, the recognition of objects and the resolution of mathematical problems, in addition to allowing more fluid and natural conversations.
In the race to offer a more advanced service to users, one of the approaches is to accelerate the general perception that we are interacting, not with a computer or a robot, but with a person, a friend even, always willing to help; It is about subtracting the “artificial” the artificial intelligence of the assistant. For many companies, overcoming the voice barrier seems to be the next logical step in the progressive humanization of the machine.
#Google #incorporate #Chirp #Vartex #platform #speak #languages