OpenAI to Start Launching Latest Iteration of Artificial Intelligence Chatbot GPT-4o | Science & Technology Updates
The latest iteration of the ChatGPT AI chatbot has been introduced and promises almost instant results in text, vision, and audio, as stated by its creator.
OpenAI mentioned that it excels in understanding visuals and sounds compared to previous versions.
It presents the opportunity for real-time ‘conversations’ with the chatbot, even allowing interruptions in its responses.
The company states that it can process any combination of text, audio, and images as input and produce outputs in various combinations of text, audio, and images.
GPT-4o will be released over the next few weeks in the midst of a competition among tech companies to advance their artificial intelligence capabilities.
The announcement on Monday showcased tasks like real-time language translation, using vision to solve a math problem on paper, and assisting a visually impaired individual in navigating London.
GPT-4o can respond to audio in as little as 232 milliseconds, averaging around 320 milliseconds, a response time comparable to that of humans, according to the company.
To address concerns regarding bias, fairness, and misinformation, the Microsoft-sponsored company reports that the new version underwent extensive testing by 70 external experts.
This development follows a public relations misstep by Google earlier this year involving images generated by its Gemini AI system.
The GPT-4o model will be available for free, with premium ‘Plus’ users enjoying a higher capacity limit for messages.
Past versions of the chatbot have raised concerns in educational institutions where students have used it to cheat by producing convincing essays.
At its launch two years ago, ChatGPT was recognized as the fastest-growing app to reach 100 million active monthly users.
This announcement also preempts Google, which is set to reveal its own new AI features at its annual developers’ conference tomorrow.