- OpenAI had demoed live vision capabilities in the Advanced Voice Mode within ChatGPT but had not shared a release timeline beyond the alpha version.
- We spotted strings that suggest that the feature, which could be called “Live camera,” could soon be released in ChatGPT’s beta version.
Many people heavily rely on ChatGPT for their daily professional and personal needs. OpenAI added a level of friendliness to ChatGPT with features like Advanced Voice Mode for natural conversations, but users have been waiting for the promised vision capabilities to also roll out. There’s good news on this front, as ChatGPT’s Live Video features in the Advanced Voice Mode could soon be rolling out to more users.
When OpenAI announced GPT-4o in May 2024, it boasted of advanced live vision capabilities coming to ChatGPT’s Advanced Voice Mode. The company famously showed off this demo where the new Advanced Voice Mode easily and seamlessly recognized the subject in the camera feed as a dog, remembered its name, recognized the ball, and associated the ball and the dog through an activity like fetch.