OpenAI has announced its latest artificial intelligence (AI) Large Language Model (LLM) GPT-4o (Here, o stands for “Omni”). The latest product can observe human emotions as well as respond in a more humanlike form better than before.
On Monday, May 13, the company demonstrated that the GPT-4o can serve as a user’s digital assistant with an enhanced ability to engage in real-time human-like conversation. The model can even interact with a vision that “reads” screenshots, photos, documents, and chats users upload.
The GPT-4o succeeds the previous GPT-4 LLM launched back in 2023. It’s said to enhance the use of ChatGPT and make it much easier. Likewise, early users will have free access to OpenAI’s latest and most advanced AI tool.
OpenAI Chief Technology Officer Mira Murati said the updated version of ChatGPT has memory capabilities. While it’s not entirely a novelty, we understand much efficient brain of the AI. As per the company, ChatGPT can now learn from previous conversations and provide real-time translation.
“GPT-4o reasons across voice, text and vision,” Murati stated during a live stream presentation of GPT-4o in San Francisco on Monday (May 13, 2024). “And this is incredibly important, because we’re looking at the future of interaction between ourselves and machines.”
OpenAI CTO Mira Murati on GPT-4o
“This is the first time that we are really making a huge step forward when it comes to the ease of use,” Murati said during the live demonstration of the product. “This interaction becomes much more natural and far, far easier.”
GPT-4o brings humanlike voice interaction
The voice mode of the GPT-4o, a hallmark of the GPT LLM is now able to provide even more humanlike suggestions to math equations and instructions rather than simply solving or answering. As evident from the live demo, the model can interpret computer code and translate languages while also interpreting the emotions of a human being.
Check out: Nepal ranks second in driving ChatGPT growth on Google Search
OpenAI teases ahead of Google’s I/O
OpenAI was rumored to introduce an AI-powered search engine to rival Google Search just ahead of the latter’s I/O. Though that didn’t happen, it still managed to launch a superior LLM. Since launching ChatGPT in November 2023, OpenAI has been a competitor for Google. For its own part, Google and Meta are also developing their own language models-backed chatbots.
In terms of Google’s I/O Developer Conference 2024, the company is expected to unveil its latest reinforcements on the Gemini AI model. Apple, likewise, is holding its own Worldwide Developers Conference in April so we can fancy the phone maker to make some AI announcements.
During the live demo of GPT-4o, the OpenAI executives demonstrated how ChatGPT could engage in real-time conversations to solve a math problem. The chatbot also successfully told a bedtime story and offered coding advice. The chatbot showed that it can now speak in a natural human-like voice and also emulate a robot voice. The AI tool was also able to describe a chart.
GPT-4o can detect users’ emotions
The GPT-4o is advanced and that translates into a more human-like dimension. The model can detect users’ emotions. During the demo, the model remarked on an executive’s breathing and told him to “calm down. ChatGPT was also able to have a conversation in multiple languages by translating and responding automatically. The tool now supports more than 50 languages, according to OpenAI.
“The new voice (and video) mode is the best computer interface I’ve ever used,” OpenAI CEO Sam Altman said in a blog post following the announcement. “It feels like AI from the movies, and it’s still a bit surprising to me that it’s real. Getting to human-level response times and expressiveness turns out to be a big change.”
Also read: Apple to adopt RCS Chat: Android-iPhone messaging to become better and easier
50 languages supported
The GPT-4o supports around 50 languages making it a highly multilingual utility. OpenAI says that it’s twice as fast in OpenAI’s API and Microsoft’s Azure OpenAI Service. And while voice compatibility is a promising aspect, it’s not a part of it yet due to the fears of misuse. But still, audio capabilities will come to “a group of trusted partners in the coming weeks,” the company maintains.
GPT-4o availability release date price
OpenAI seeks to roll out GPT-4o in the coming months. Non-paid ChatGPT users will enjoy a limited number of interactions with the new GPT-4o model before automatically reverting to the GPT-3.5 model. On the other hand, paid users will have wholesome privileges with the ChatGPT chatbot. According to Murati ChatGPT desktop app will be launched with the GPT-4o capabilities. Additionally, the model will be available to developers who would be eager to design their own custom chatbots from OpenAI’s GPT store, a feature that will be issued also to non-paying users.
It’s also noteworthy that the GPT-4o comes free in the free tier ChatGPT use and for OpenAI’s premium ChatGPT Plus and Team plans with 5x higher message limits. But the ChatGPT reverts back to GPT-3.5 when users reach the message rate limit. The voice experience could be available to Alpha for Plus users in June.
And let’s not rule out the latest model to come to Microsoft. The tech giant has invested heavily to integrate its AI technology into its own products including Bing Search. As with the global users, GPT-4o powered-ChatGPT will be available in Nepal in the next few months. OpenAI will soon release a fresh ChatGPT UI for the web with a more user-friendly home screen and message layout.
OpenAI said that currently over 100 people are using ChatGPT and the latest GPT-4o will give more people a reason to use its technology. The new AI product comes just when Google and Meta are also bolstering their AI game. The race perhaps has now reached its next phase with OpenAI’s new LLM launch.