OpenAI has announced an exciting new feature for its chatbot ChatGPT, enabling it to process and engage with users based on real-time observations from video feeds. This capability was first hinted at seven months ago and has now been officially unveiled during a livestreamed event.
With the latest update, ChatGPT can utilize a smartphone’s camera to recognize objects and converse about what it sees on the screen. For instance, users can seek assistance in crafting responses in messaging apps or receive step-by-step guidance for tasks like brewing coffee, all while ChatGPT responds dynamically based on its visual input.
This video interaction feature will start rolling out to paid ChatGPT Plus and Pro subscribers immediately, with a broader introduction for enterprise and educational users set for January.
Since launching ChatGPT two years ago, OpenAI has spearheaded advancements in text-based chatbots, driving a trend towards multimodal functionalities that can respond to audio, images, and videos. This move promises to make digital assistants more engaging and versatile.
The announcement marks just one of several developments as OpenAI embarks on a 12-day livestream series to showcase new products. This series also includes the introduction of a pricier ChatGPT Pro subscription and the rollout of an AI video generation tool called Sora.
In summary, OpenAI continues to push the boundaries of chatbot technology, integrating real-time video processing to enhance user interactions and support various tasks, heralding a more interactive and adaptive future for AI technology. This innovation highlights the continuous evolution in how we engage with digital tools, promising greater convenience and functionality in our daily lives.