A new version of ChatGPT can read facial expressions, mimic human voice patterns and have near real-time conversations, its creators have revealed.
OpenAI demonstrated the upcoming version of the artificial intelligence (AI) chatbot, called GPT-4o, in an apparently real-time presentation on Monday (May 13). The chatbot, which spoke out loud with presenters through a phone, appeared to have an eerie command of human conversation and its subtle emotional cues — switching between robotic and singing voices upon command, adapting to interruptions and visually processing the facial expressions and surroundings of its conversational partners. During the demonstration, the AI voice assistant showcased its skills by completing tasks such as real-time language translation, solving a math equation written on a piece of paper and guiding a blind person around London’s streets.