OpenAI will soon be releasing a cheaper, lighter model called GPT-4o Mini for developers. It costs significantly less than full-sized models and is said to be more capable than GPT-3.5.
Building apps using the company’s models can generate a hefty bill. Developers without the means to afford to tinker with it can get priced out of it entirely and might opt for cheaper models like Google’s Gemini 1.5 Flash or Anthropic’s Claude 3 Haiku. Now, the company is entering the light model game.
“I think GPT-4o Mini really gets at the OpenAI mission of making AI more broadly accessible to people. If we want AI to benefit every corner of the world, every industry, every application, we have to make AI much more affordable,” Olivier Godement, who leads the API platform product, told The Verge.
Also Read: Realme Watch S2 Powered By ChatGPT To Launch On July 30
Who Can Use the GPT-4o Mini Model?
ChatGPT users on Free, Plus, and Team plans will be able to use GPT-4o Mini instead of GPT-3.5 Turbo, with Enterprise users getting access next week. This means GPT-3.5 will no longer be an option for ChatGPT users, but it will still be available for developers through the API if they prefer not to switch to GPT-40 Mini. Godement said GPT-3.5 will get retired from the API at some point.
The new model will also bring support for text and vision in the API, and the company says it will soon be handling multimodal inputs and outputs like audio and video. With all these capabilities, this could look like more capable virtual assistants that can understand users’ travel itineraries and create suggestions.
Also Read: Apple Warns Millions Of iPhone Users Against Google Chrome
Is GPT-4o Mini Better Than Precursor?
The new model was amanged to achieve an 82 percent score on the Measuring Massive Multitask Language Understanding (MMLU), a benchmark exam consisting of about 16,000 multiple-choice questions across 57 academic subjects. When the MMLU was first launched in 2020, most models weren’t that great, which was the goal since the models had gotten too advanced for previous benchmark exams.
GPT-3.5 managed to score 70 percent on this benchmark, GPT-4o scored 88.7 percent, and Google claims Gemini Ultra to have the highest-ever score of 90 percent. In comparison, the competing models Claude 3 Haiku and Gemini 1.5 Flash scored 75.2 percent and 78.9 percent, respectively.
It’s imperative to note that researchers are aware of benchmark tests like the MMLU, as how it’s administered differs slightly from company to company. That makes different models’ scores difficult to compare, as per The New York Times report.