ChatGPT’s Vision feature has garnered a lot of attention since launch of the feature on September 25. OpenAI is using the multimodal abilities of GPT-3.5 and GPT-4 in order to power the Image understanding of ChatGPT. Users can now upload one or more images to ask ChatGPT questions like explore the contents of my fridge to plan a meal, or analyze a complex graph for work-related data.
ChatGPT vision feature is currently only available to the company’s Plus and Enterprise users and OpenAI has promised to make it available for developers in the coming weeks.
ChatGPT can now talk to you. Here’s how to use the newly released features by OpenAI
Also Read : Vise Intelligence wants to use AI to assist financial advisors
5 Ways users have taken advantage of ChatGPT’s vision capabilities:
1) Helping the learning process:
A user leveraged ChatGPT’s vision capabilities to help understand the diagram of a human cell, pointing out a potential use case for the chatbot in the education sector.
2) Help understand complex messages:
Another user asked ChatGPT to help understand the meaning of an image and the chatbot complied by giving a point-by-point explanation on the topic.
One user asked ChatGPT to identify the scene that an image belonged to and OpenAI ‘s chatbot gave out complete information about the scene from the movie ‘Gladiator’ and even went on to reiterate what the character said in that particular scene.