I’ve been using ChatGPT’s advanced voice model since it was activated on my Plus account, a few months after Open AI was officially introduced. So far I’ve had to do the VPN trick with the response delays added by the tunneled connection. But that’s no longer necessary: ChatGPT’s advanced voice assistant is officially live in Europe.
The parallel between each movie and ChatGPT’s advanced voice assistant is not accidental; Open AI itself was (overly) inspired by Spike Jonze’s famous movie. And since I’ve tried this I can only agree, it’s easy to forget that we’re talking to a machine. Reflections of sound, laughter, breathing… The magic of details is what reinforces the human experience. Since I no longer have to use a VPN, I can put ChatGPT as my personal assistant.
ChatGPT’s advanced voice model now works in Europe
As I said, I have been using and chatting with the new voice model since it became active on my ChatGPT Plus at the end of September. To use this, I had to activate my VPN beforehand so that the IP address pointed to the United States. The magician would then switch from the old environment, which was already very good, to the new environment, which was perfect. As Open AI itself has announced, this is no longer necessary.
What makes advanced AI great is that the speech mode has almost no delay (responds like a person would after listening to you), Open AI models perfectly adapt the language to the request, and most importantly, the voice creates a voice. personality It’s almost human in the details. It has a certain mechanical tone at first, but it can be changed at will. After my tests, the accent that felt most natural to me was the Andalusian accent.
The advanced sound model has nine voices. My favorite is Vale with an Andalusian accent: the best natural recipe I’ve found.
It is a matter of curiosity that the voice assistant depends on the selected language model. For example, if I choose o1-preview the argument becomes much more complex, but it takes a few seconds to think about it; This causes speech impediment. It’s a little faster with the o1-mini but still sounds very similar to the old one. GPT-4o is excellent.
I can ask it to explain any question, start a random conversation, ChatGPT allows me to model how audio behaves on request, helps me with documentation tasks when I can’t take my eyes off the keyboard, and, less importantly, it amuses me. Because he has great mastery.
I was able to chat with the AI by impersonating it so it could tell me jokes, crack stories, it’s great for practicing English (sometimes I ask it for a “talking” session so I don’t get rusty), and I put it to the test on two phones at once to see how far it got in conversation. Once an entire language was invented He adds new words by talking to himself. It amazes me how you create a story by launching mobile-to-mobile challenges.
ChatGPT’s enhanced audio is very good but far from perfect
Open AI focused on naturalness, I have previously emphasized that the details in the voice are what provides humanity. Language models do their own part in analyzing and responding The voice speaks what the artificial intelligence wrote. It’s great to be able to interact with ChatGPT without looking or touching your phone. It’s fine now.
New sound model although it works perfectly Far from what Open AI promises: It cannot analyze the environment with the camera, does not perform mathematical analysis when focusing on a problem with the mobile phone, and generally does not allow interaction with the environment as if it had virtual eyes. Everything about the camera still doesn’t work, which reduces the huge promised potential.
Open AI has yet to incorporate all interactions with the real world into its audio model, taking advantage of the mobile camera.
Another point is that although I can use ChatGPT as a personal assistant, it has zero ability to manage my mobile phone or connected objects. I wish I could tell him to turn off the Wi-Fi, disconnect my house alarm, or check the cameras to tell me if he sees anything strange, but no. Although technically feasible, I doubt Open AI will introduce this in the future: The application can interact with mobile hardware and software using Android APIs.
Cover image | DALL-E 3 hosted on ChatGPT Plus
Xataka on Android | All phones updated to Android 15 and when they will start updating
Xataka on Android | How to share your Android data connection with other devices