OpenAI has enhanced its chatbot ChatGPT by introducing the capability to process and discuss real-time observations from video feeds, a feature it had initially hinted at seven months ago. This development was announced during a livestreamed event on Thursday.
With this new functionality, ChatGPT can now recognize objects through a smartphone camera and verbally interact with users based on what it sees on-screen. For instance, users can request assistance in replying to messages in an app or seek live guidance for tasks like making coffee.
The video feature will start rolling out on Thursday for ChatGPT Plus and Pro subscribers, while enterprise and educational users will gain access in January.
Since the launch of ChatGPT two years ago, OpenAI has been at the forefront of innovations in text-based chatbots. The company and its competitors have increasingly incorporated multimodal features that allow their services to respond to audio, images, and video, thus creating a more interactive digital assistant.
This announcement is part of a series of product unveilings that OpenAI plans to broadcast over the next 12 days. The company is also launching a more expensive ChatGPT Pro subscription and a new AI video generation tool called Sora.
In summary, with these advancements, users can look forward to a more engaging and interactive experience with ChatGPT, enhancing how they interact with technology in their daily lives. This leap in functionality highlights OpenAI’s commitment to pushing the boundaries of AI and improving user experience. The continued investment in multimodal capabilities offers exciting possibilities for the future of digital assistants.