OpenAI, the company behind the ChatGPT artificial intelligence engine, is about to introduce a new model of multimodal artificial intelligence capable of interacting through dialogue and recognizing objects, and designed specifically for use on smartphones. Several insiders say that this could be one of the announcements planned for the presentation that the company will broadcast online on Monday, at 7pm Italian time. The new assistant promises faster and more accurate interpretation of images and audio than previous transcription and speech synthesis models offered separately by the company. It will be especially useful for customer service operators, helping them better understand intonation in callers’ voices. The model could also assist students or translate signs into the real world.
Sources claim that this model surpasses GPT-4 Turbo in answering certain types of questions. In parallel, it seems that OpenAI is preparing a new ChatGPT feature for making phone calls. This hypothesis arises from a screenshot published by developer Ananay Arora, which shows codes related to calling functionality. Arora also noted that OpenAI has configured servers intended for real-time audio and video communication. Despite the speculation, OpenAI CEO Sam Altman denied that the upcoming announcement directly concerns GPT-5, which will arrive later this year. Altman also ruled out that the announcement includes a new AI-powered search engine. The timing of these developments is crucial, as Google is testing the use of artificial intelligence to make phone calls and there are rumors of a project called “Pixie”, a multimodal replacement for the Google Assistant capable of recognizing objects through the camera a device.
#ChatGPT #OpenAI #challenges #Google #bring #phone #calls