OpenAI is likely just days away from the launch of GPT-5, the latest installment of the artificial intelligence technology that powered the ChatGPT juggernaut in 2022 and users will be able to scrutinize if the step up from GPT-4 is on par with the research lab's previous improvements.
The AI giant, which claimed earlier this week it is "on track" to 700 million weekly active users, also unveiled on Tuesday two new "open" models that can be customized and run locally on a laptop.
Dubbed "gpt-oss-120b" and "gpt-oss-20b," the two models are so-called "open-weight language models" that allow users to access the AI's parameters and customize them, in a departure from the company's ChatGPT bot, which is based on a "closed" model.
"Developed with open-source community feedback, these models deliver meaningful advancements in both reasoning capabilities & safety," OpenAI said in a statement on X.
Parallel with this, media reports indicated that GPT-5 is coming "early" in August, with the company CEO, Sam Altman, also saying in July that they would be releasing it "soon."
"We are releasing GPT-5 soon but want to set accurate expectations: this is an experimental model that incorporates new research techniques we will use in future models. We think you will love GPT-5, but we don't plan to release a model with IMO gold level of capability for many months," he wrote on X.
Two early testers of the new model told Reuters they have been impressed with its ability to code and solve science and math problems, but they believe the leap from GPT-4 to GPT-5 is not as large as the one from GPT-3 to GPT-4. The testers, who have signed non-disclosure agreements (NDAs), declined to be named for this story.
OpenAI declined to comment to Reuters.
GPT-4’s leap was based on more compute power and data, and the company was hoping that "scaling up" in a similar way would consistently lead to improved AI models.
But OpenAI, which is backed by Microsoft and is currently valued at $300 billion, ran into issues scaling up. One problem was the data wall the company ran into, and OpenAI's former chief scientist, Ilya Sutskever, said last year that while processing power was growing, the amount of data was not.
He was referring to the fact that large language models (LLMs) are trained on massive datasets that scrape the entire internet, and AI labs have no other options for large troves of human-generated textual data.
Apart from the lack of data, another problem was that "training runs" for large models are more likely to have hardware-induced failures given how complicated the system is, and researchers may not know the eventual performance of the models until the end of the run, which can take months.
OpenAI has not said when GPT-5 will be released, but the industry expects it to be any day now, according to media reports.
Boris Power, head of Applied Research at OpenAI, said in an X post on Monday: "Excited to see how the public receives GPT-5."
"OpenAI made such a great leap from GPT-3 to GPT-4, that ever since then, there has been an enormous amount of anticipation over GPT-5,” said Navin Chaddha, managing partner at venture capital fund Mayfield, who invests in AI companies but is not an OpenAI investor. "The hope is that GPT-5 will unlock AI applications that move beyond chat into fully autonomous task execution."
Nearly three years ago, ChatGPT introduced the world to generative AI, dazzling users with its ability to write humanlike prose and poetry, quickly becoming one of the fastest-growing apps ever.
In March 2023, OpenAI followed up ChatGPT with the release of GPT-4, a large language model that made huge leaps forward in intelligence. While GPT-3.5, an earlier version of the model, received a bar exam score in the bottom 10%, GPT-4 passed the simulated bar exam in the top 10%.
GPT-4 then became the model to beat and the world came to terms with the fact that AI models could outperform humans in many tasks.
Soon, other companies were catching on. The same year, Alphabet's Google and Anthropic – which is backed by Amazon and Google – released competitive models to GPT-4. Within a year, open-source models on par with the GPT-4, such as Meta Platforms' Llama 3 models, were released.
Along with training large models, OpenAI has now invested in another route, called "test-time compute," which channels more processing power to solve challenging tasks such as math or complex operations that demand human-like reasoning and decision-making.
Altman said earlier this year that GPT-5 would combine both test-time compute and its large models. He also said that OpenAI's model and product offerings had become "complicated."