The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved
Curated by THEOUTPOST
On Fri, 19 Jul, 12:03 AM UTC
28 Sources
[1]
OpenAI Unveils Cheaper Small AI Model GPT-4o Mini
(Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said. (Reporting by Deborah Sophia in Bengaluru; Editing by Vijay Kishore)
[2]
OpenAI unveils cheaper small AI model GPT-4o mini
July 18 (Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said. (Reporting by Deborah Sophia in Bengaluru; Editing by Vijay Kishore)
[3]
OpenAI unveils cheaper small AI model GPT-4o mini
July 18 (Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy-intensive, allowing the startup to target a broader pool of customers. Microsoft-backed (MSFT.O), opens new tab OpenAI, the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals such as Meta (META.O), opens new tab and Google (GOOGL.O), opens new tab rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. With the mini model currently supporting text and vision in the application programming interface, OpenAI said support for text, image, video and audio inputs and outputs would be made available in the future. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini -- which has knowledge up to October 2023 -- starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said. Reporting by Deborah Sophia in Bengaluru; Editing by Vijay Kishore Our Standards: The Thomson Reuters Trust Principles., opens new tab
[4]
OpenAI Announces GPT-4o Mini: Cheaper And Faster AI Chatbot For More People
OpenAI wants more people to pay for ChatGPT and a cheaper model makes sense (Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said.
[5]
OpenAI unveils cheaper small AI model GPT-4o mini
MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage.ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy-intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals such as Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. With the mini model currently supporting text and vision in the application programming interface, OpenAI said support for text, image, video and audio inputs and outputs would be made available in the future. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini - which has knowledge up to October 2023 - starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said.
[6]
OpenAI released a smaller, cheaper version of its most powerful AI model
OpenAI introduced a smaller, more cost-efficient version of its most powerful model GPT-4o -- and it could mean even more AI-powered apps. The new model, GPT-4o mini, is smarter and more than 60% cheaper than GPT-3.5 Turbo, according to OpenAI, meaning it is more affordable for developers building applications on OpenAI's models. GPT-4o mini is priced at 15 cents per 1 million input tokens -- or the unit of information an AI model uses, for example, a word or phrase -- and 60 cents per 1 million output tokens. That is equivalent to 2,500 pages in a book, OpenAI said. "We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable," OpenAI said in a statement. On the Massive Multitask Language Understanding (MMLU) benchmark, which measures language models' capabilities across 57 subjects including math and law, GPT-4o mini scored 82%, outperforming other small models and GPT-3.5 Turbo, according to OpenAI. The small model will replace GPT-3.5 Turbo in ChatGPT, and Free, Plus, and Team users have access to the new model now. At its release, GPT-4o mini can support text and vision in the API, or application programming interface, which allows developers to build their own AI models off of OpenAI's technology. In the future, GPT-4o will have support for text, image, video, and audio, OpenAI said. The small model has knowledge up to October 2023. "We envision a future where models become seamlessly integrated in every app and on every website," OpenAI said. "GPT-4o mini is paving the way for developers to build and scale powerful AI applications more efficiently and affordably." OpenAI unveiled ChatGPT-4o, in May, which is multimodal, meaning it can see, hear, and have real-time conversations.
[7]
OpenAI reveals cheaper mini version of its flagship GPT-4o
OpenAI is advertising GPT-4o mini as a model that can perform various tasks at a low cost, as the AI sector continues to shift into developing small language models. OpenAI has released a more cost-efficient version of its flagship GPT-4o model, to reach a wider audience and expand how its products are used. The company revealed its major GPT-4o upgrade earlier this year and said it will bring advanced features to its AI products, such as giving ChatGPT the ability to respond faster to text, audio and images - though some upgrades are facing delays. But instead of power and performance, OpenAI is advertising GPT-4o mini as a model that can perform various tasks at a low cost and latency. The company also claims it is an upgrade from its previous "small models" such as GPT-3.5 Turbo. Users of ChatGPT's Free, Plus and Team are now able to access GPT-4o mini. "We envision a future where models become seamlessly integrated in every app and on every website," OpenAI said in a blogpost. "GPT-4o mini is paving the way for developers to build and scale powerful AI applications more efficiently and affordably." OpenAI said it tested GPT-4o mini on "several key benchmarks" and claimed its model surpassed various other small models on the market "across both textual intelligence and multimodal reasoning". However, it's important to take company claims about its own models with a pinch of salt. Last year, Google-owned DeepMind claimed that one of its AI models found 2.2m new crystals that could potentially be used to create new materials. This claim was later criticised by a group of researchers who said there was "scant evidence" the DeepMind AI found for compounds that "fulfil the trifecta of novelty, credibility, and utility". Earlier this year, the AI Index claimed that robust evaluations for large language models are "seriously lacking" and there is a lack standardisation in responsible AI reporting. Large language models became the golden egg of the tech sector after the rise of ChatGPT, prompting many companies to develop their own models and applications around generative AI. But these models also have issues when it comes to both cost and energy use. As a result, some of the key players in the AI sector have been creating smaller, more accessible models. Google and Microsoft - two of the biggest players in the AI sector currently - have both made moves towards more lightweight AI options for their customers this year. Find out how emerging tech trends are transforming tomorrow with our new podcast, Future Human: The Series. Listen now on Spotify, on Apple or wherever you get your podcasts.
[8]
OpenAI Launches ChatGPT-4o Mini: An Affordable AI Model Designed For Developers
OpenAI's ChatGPT-4o Mini unveiled as a lightweight solution for developers OpenAI has once again set the AI world abuzz with its latest creation, the GPT-4o Mini. Introduced today under the leadership of Sam Altman, this new model marks a significant stride forward in making advanced AI more accessible. Designed as a more budget-friendly option compared to its larger counterparts, GPT-4o Mini aims to democratize AI by offering powerful capabilities at a lower cost. Standing out for its improved performance and efficiency over predecessors like GPT-3.5, GPT-4o Mini is celebrated by OpenAI as their most cost-effective small model to date. It's designed to open up new possibilities across various fields of application, as highlighted in OpenAI's official blog, where it achieved an impressive 82% on the MMLU benchmark, surpassing even GPT-4 in certain areas. One of GPT-4o Mini's standout features is its versatility. It handles a wide array of tasks with minimal delay, from managing multiple model requests to engaging in real-time interactions like customer support chatbots. Initially supporting text and vision via its API, future updates will extend its compatibility to include inputs and outputs across text, images, videos, and audio. In technical terms, GPT-4o Mini operates with a context window spanning 128K tokens and can generate outputs up to 16K tokens long. It retains knowledge up to October 2023 and incorporates enhanced tokenization methods, making it particularly adept at processing non-English text efficiently. Performance benchmarks underscore its superiority over previous models like GPT-3.5 Turbo, especially in tasks involving multimodal reasoning and textual intelligence. Furthermore, GPT-4o Mini inherits the robust safety measures of its predecessor, GPT-4o, meticulously evaluated through human and automated assessments involving over 70 experts from diverse fields. This ensures the model meets stringent safety and reliability standards, reinforcing OpenAI's commitment to ethical AI development. Developers are enthusiastic about GPT-4o Mini's enhanced capabilities, which include better handling of function calls and extended context comprehension compared to earlier iterations. This advancement empowers developers to create sophisticated applications that seamlessly integrate with external systems, delivering precise and swift results. As OpenAI continues to push the boundaries of AI accessibility and performance, the launch of GPT-4o Mini represents a milestone in the evolution of artificial intelligence. With its affordability, versatility, and robust performance metrics, GPT-4o Mini is set to drive innovation across industries, offering new avenues for businesses and developers to harness AI in transformative ways. SEE ALSO: Apple iPhone 13, MacBook Air M1, OnePlus 12 Receive Massive Price Cut During Amazon Prime Day 2024
[9]
OpenAI offers GPT-4o mini to slash the cost of applications
OpenAI on Thursday unveiled a stripped-down version of its GPT-4o large language model, GPT-4o mini, which it said has better accuracy than GPT-4 on tasks, and costs dramatically less than GPT-3.5 "Turbo" when used by developers, which it said can boost the construction of applications that use the AI model extensively. The company touts the new AI model as "the most cost-efficient small model in the market," although, as with most OpenAI releases, no technical details are available about GPT-4o mini, such as the number of parameters, hence, it's unclear what "small" means in this case. (An "AI model" is the part of an AI program that contains numerous neural net parameters and activation functions that are the key elements for how an AI program functions.) Also: How to use ChatGPT to create an app GPT-4o mini "is priced at 15 cents per million input tokens and 60 cents per million output tokens, an order of magnitude more affordable than previous frontier models and more than 60% cheaper than GPT-3.5 Turbo," said OpenAI in a blog post emailed to ZDNET. That reduction in cost, said the company, will aid the development of applications that are affected by volume of activity. For example, applications that must make multiple API (application programming interface) calls, or that use larger "context windows" to retrieve materials (say, to retrieve an entire code-base when developing an app), or that have to interact frequently with the end user, such as a help desk support bot, will benefit from the reduction in per-transaction cost, said OpenAI. The model, says OpenAI, outperforms the standard GPT-4 model when used as a chatbot, based on crowd-sourced tests by the Lmsys leaderboard. It also "surpasses GPT-3.5 Turbo and other small models on academic benchmarks across both textual intelligence and multimodal reasoning," and supports as many languages as the standard GPT-4o model. The new model is available immediately to developers via the Assistants API, Chat Completions API, and Batch API, and can be used instead of GPT-3.5 Turbo in ChatGPT's free, plus, and team accounts. The model offers only text and image support at the moment, with audio and video to be added at an unspecified date. The GPT-4o mini context window is 128,000 tokens, and its training data is current through October of 2023.
[10]
OpenAI Previews Slimmed-Down and Cheaper GPT-4o Mini Model
OpenAI is launching GPT-4o Mini, a scaled-down and cheaper version of newest GPT-4o model. GPT-4o Mini can accept and output text and image content; over time, it will be able to process other types of requests, Bloomberg reports. It replaces GPT-3.5 Turbo in ChatGPT and is available to those using the free and paid versions of OpenAI's chatbot. Enterprise users get it next week. For developers, GPT-4o Mini will be a more affordable way to use OpenAI's tech. They'll pay $0.15 per million input tokens and $0.60 per million output tokens -- 60% cheaper than GPT-3.5 Turbo, TechCrunch reports. It has a context window of 128,000 tokens and is trained on data up to October 2023. At I/O in May, Google also launched a smaller AI model "optimized for narrower or high-frequency tasks," dubbed Gemini 1.5 Flash. At OpenAI, the company is aiming for a better multi-modal experience. "If you think about the way we as humans process the world and engage with the world, we see things, we hear things, we say things -- the world is much bigger than text," OpenAI COO Brad Lightcap tells CNBC. OpenAI launched GPT-4o earlier this year with promises of it being the most humanlike AI yet thanks to features like Voice Mode, which is currently delayed while OpenAI works out some kinks in the code. Look for it later this year.
[11]
OpenAI's GPT-4o mini launched | Smaller and cheaper than GPT-3.5 Turbo model
OpenAI announced the release of GPT-4o mini, which it called its "most cost-efficient small model." GPT-4o mini can support text and vision in the API, while support for text, image, video and audio inputs and outputs is yet to come. Per the ChatGPT-maker, GPT-4o mini has a context window of 128K tokens and supports up to 16K output tokens per request. Its knowledge cut-off is October 2023 and the model handles non-English text in a more cost-effective way, claimed the company. While the name might be "mini," OpenAI stressed that the small model could hold its own against both smaller rivals as well as provide an experience comparable to larger ones. (For top technology news of the day, subscribe to our tech newsletter Today's Cache) AI accessibility? Blind gamer puts ChatGPT to the test "GPT-4o mini surpasses GPT-3.5 Turbo and other small models on academic benchmarks across both textual intelligence and multimodal reasoning, and supports the same range of languages as GPT-4o," said OpenAI. ChatGPT's Free, Plus, and Team users can access the new model immediately, while Enterprise users will get access from next week. OpenAI noted that safety measures were in place from the pre-training stage so that the model would not learn from hate speech, adult content, sites that primarily aggregate personal information, and spam. In addition, the model has been fortified to better stand against jailbreak attempts, prompt injections, and system prompt extractions. "GPT-4o mini surpasses GPT-3.5 Turbo and other small models on academic benchmarks across both textual intelligence and multimodal reasoning, and supports the same range of languages as GPT-4o. It also demonstrates strong performance in function calling, which can enable developers to build applications that fetch data or take actions with external systems, and improved long-context performance compared to GPT-3.5 Turbo," said OpenAI in its statement introducing the new model. The AI company backed by Microsoft was criticised by whistleblowers and former employees who claimed that it did not take enough safety precautions when releasing new products, and that it tried to stop employees from speaking up about the same. Read Comments
[12]
OpenAI Introduced A New Small AI Model, ChatGPT-4o mini That Is Faster, Cost Efficient, And Could Outperform Others
Microsoft's OpenAI has undoubtedly been leading the AI space with its cutting-edge technological advancements and pioneered exploration of the technology's untapped potential. The company keeps bringing new, more enhanced versions of ChatGPT that are better equipped and have greater capabilities. The latest iteration was the GPT-4 turbo, which was said to upgrade performance and improve cost efficiency. The platform keeps evolving, and on Thursday, OpenAI announced a new small AI model that will outperform the current versions and is available for developers and consumers to access. OpenAI is focused heavily on making ChatGPT's experience comprehensive, taking on other AI models by the competition by bringing in more enhanced capabilities with every version rolled out. To continue improving, the company rolled out a new version that will be lightweight and economical so users can get true value for money. The latest small AI model is called the GPT-4o mini. It is said to offer more advanced performance and require less power, saving on resources and targeting a wider consumer base with its affordable cost. This new model would replace GPT-3.5 Turbo for the smallest AI model yet. Developers wanting to use OpenAI models for building apps earlier incurred a higher cost, so the light models introduced by other companies, such as Google's Gemini 1.5 Flash, gained more popularity. By bringing a smaller, lighter model, OpenAI is stepping up its game, which could outperform all other AI models and be popular. OpenAI's claim that the model's performance surpasses existing models is backed up by a comparison of different models drawn by Artificial Analysis. The GPT-4o mini scored 82 percent on Massive Multitask Language Understanding (MMLU), signifying the model's ability to understand different contexts and expand over wider domains. In comparison, Google's Gemini Flash scored 77.9 percent, and Anthropic's Claude Haiku scored 73.8 percent, significantly below the new model's multitasking abilities. Since smaller AI models rely on less power, they tend to be more energy efficient and save on costs. Due to limited computational resources being used, they are designed to be more efficient and offer advanced performance. OpenAI, by developing its own compact and light model, could offer robust capabilities and the ability for the model to perform complex tasks more effectively. OpenAI's claim that GPT-4o might outperform leading smaller language models by being more cost-efficient, smarter, and faster does seem promising, given its plans to add audio and video support later on. ChatGPT Free, Plus, and Team users can start using the new model, and enterprise users will be given access next week.
[13]
OpenAI Now Has a GPT-4o Mini. Here's Why That Matters
ChatGPT maker OpenAI introduced a smaller model called GPT-4o Mini on Thursday, which it says is smarter and cheaper than GPT-3.5 Turbo, an earlier model that was built for simple tasks like dialogue. OpenAI hopes developers will use GPT-4o Mini to "significantly expand the range of applications built with AI," according to a blog post. Chatbots like ChatGPT are the interface we use to communicate with large language models, or LLMs, like GPT-4o Mini and the original, much larger GPT-4o. These models are trained to understand how we use language so they can generate content that sounds human. An LLM can have 1 billion or more parameters, which is a measure of how much content it can ingest before producing a response to your prompt. That means LLMs can learn from and understand a lot, but they aren't ideal for every situation. They can be expensive and consume a lot of energy because of the need for expansive server farms and access across the cloud. A small language model is a compromise of sorts. It offers AI horsepower and speed but doesn't require the same computing resources or cost. Microsoft's Phi-3 Mini, which is built to run on phones and PCs, is one example. Google's Gemini 1.5 Flash, which is designed for high-volume, high-frequency tasks like generating captions and extracting data from forms, is another. Now we have GPT-4o Mini as well. Both free and paid ChatGPT users can access GPT-4o Mini starting Thursday in place of GPT-3.5, which was released in November 2022. GPT-4o Mini currently supports text and vision in the OpenAI API, which is what developers use to build new applications based on OpenAI technology. Support for text, image, video and audio inputs and outputs is "coming in the future," the post said. Enterprise users will have access to GPT-4o Mini starting the week of July 22. OpenAI said GPT-4o Mini excels in mathematical reasoning and coding, but has also demonstrated skills in tasks that require reasoning. Financial tech startup Ramp and email app Superhuman tested out GPT-4o Mini to extract data from files and generate email responses, according to the post. The new model has a context window of 128,000 tokens, which is a measurement of how much it can remember in a given conversation. By way of comparison, GPT-4o has the same context window, while GPT-3.5 Turbo has a context window of 16,000 tokens. GPT-4o Mini costs 15 cents per million input tokens and 60 cents per million output tokens, which OpenAI said is about equal to 2,500 pages in a book. GPT-4o, which was released in May, costs $5 per million input tokens and $2.50 per million output tokens. "We envision a future where models become seamlessly integrated in every app and on every website," the blog post said. "GPT-4o mini is paving the way for developers to build and scale powerful AI applications more efficiently and affordably."
[14]
OpenAI unveils cheaper small AI model GPT-4o mini
(Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market.
[15]
OpenAI Introduces GPT-4o Mini, 30x Cheaper than GPT-4o
Priced at 15 cents per million input tokens and 60 cents per million output tokens, GPT-4o mini is 30x cheaper than GPT-40 and 60% cheaper than GPT-3.5 Turbo. OpenAI chief Sam Altman made a cost comparison, saying, "Way back in 2022, the best model in the world was text-davinci-003. It was much, much worse than this new model. It cost 100x more." The model excels in various tasks, including text and vision, and supports a context window of 128K tokens with up to 16K output tokens per request. GPT-4o mini demonstrates superior performance on benchmarks, scoring 82% on the MMLU, 87% on MGSM for math reasoning, and 87.2% on HumanEval for coding tasks. It outperforms other small models like Gemini Flash and Claude Haiku in reasoning, math, and coding proficiency. GPT-4o mini's low cost and latency enable a wide range of applications, from customer support chatbots to API integrations. It currently supports text and vision, with future updates planned for text, image, video, and audio inputs and outputs. Safety measures are integral to GPT-4o mini, incorporating techniques like reinforcement learning with human feedback (RLHF) and the instruction hierarchy method to improve model reliability and safety. GPT-4o mini is now available in the Assistants API, Chat Completions API, and Batch API. It will be accessible to Free, Plus, and Team users in ChatGPT today, and to Enterprise users next week. Fine-tuning capabilities will be introduced soon. GPT-4o mini comes after OpenAI co-founder Andrej Karpathy recently demonstrated how the cost of training large language models (LIMs) has significantly decreased over the past five years, making it feasible to train models like GPT-2 for approximately $672 on "one 8XH100 GPU node in 24 hours". "Incredibly, the costs have come down dramatically over the past five years due to improvements in compute hardware (H100 GPUs), software (CUDA, cuBLAS, cuDNN, FlashAttention) and data quality (e.g., the FineWeb-Edu dataset)," said Karpathy. That explains how Tech Mahindra was able to build Project Indus for well under $5 million, which again, is built on GPT-2 architecture, starting from the tokeniser to the decoder. It would be interesting to see what innovative applications developers will create using this new AI model. Looks like it's already in motion. A few days back a mysterious model had appeared on the Chatbot Arena. Unsurprisingly, that model is none other than GPT-4o mini.
[16]
ChatGPT maker OpenAI goes smaller and cheaper with new AI tech
The new version of its AI model, called GPT-4o mini, is 60% cheaper to use than the model that powered ChatGPT until recently, called GPT-3.5 turbo, the company said, and scores higher on tests it uses to measure intelligence and effectiveness. OpenAI is the best-known brand in the booming space of generative artificial intelligence, so its embrace of efficiency is a notable indicator of a broader turn for the industry. Previously, OpenAI and competitors like Google and Microsoft focused primarily on developing the biggest and most powerful AI models. Now they are trying to balance those pushes with releasing small models that can be more profitable for them and more useful for business clients that only need AI tech to do certain tasks well. Google, and startups Anthropic, Mistral and Cohere have also released smaller models this year. Microsoft has played up a family of small models named Phi it said was 1/100th the size of the model behind ChatGPT at the time. Many generative AI startups, as well as new entrant Apple, are also focusing on small models that are cheaper to build because they are typically trained on smaller batches of data. They also require less computing power to operate and can in some cases work directly off hardware like a phone. OpenAI said GPT-4o mini can currently interpret text and image inputs and will later add the capability to scan and output audio and video. The technology was rolled out to people who use the free and premium versions of ChatGPT starting Thursday. The company said it would be available to corporate customers next week.
[17]
What Is GPT-4o Mini, OpenAI's Newest Launch?
Even a baby spider is cuter than its adult counterpart, so when OpenAI debuted a mini version of its most powerful model, you'll have to excuse my instinct to say "Awwww!" That's probably not always the right response, instincts be damned. Because there isn't very much about this new mini version of GPT-4o that we'd describe as "cute." OpenAI's new AI model, GPT-4o mini, is the AI startup's latest effort to make its chatbot the biggest in the game -- even if it's tiny. The company says it is its "most cost-efficient small model." "We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable," OpenAI said in a press statement. OpenAI went on to specify that the tool currently only supports "text and vision in the API," but said that "support for text, image, video and audio inputs and outputs" will be integrated in the future. It did not specify when that integration will take place. "I think GPT-4o Mini really gets at the OpenAI mission of making AI more broadly accessible to people. If we want AI to benefit every corner of the world, every industry, every application, we have to make AI much more affordable," Olivier Godement, who leads the API platform product, told The Verge. On Thursday, most ChatGPT users -- those who use it on a free, plus, or team plan -- will be able to use GPT-4o Mini instead of GPT-3.5 Turbo. Enterprise users will gain access to the mini model next week.
[18]
OpenAI Launches New GPT-4o mini AI Model; Faster and More Efficient Than GPT 3.5 Turbo | MySmartPrice
Free users of ChatGPT will now have access to higher and more accurate responses. OpenAI has launched its new AI model GPT-4o mini with improved multimodal reasoning. This new large language model (LLM) is faster and supports longer contextual conversations compared to GPT 3.5 Turbo. It will be available for both free and paid users of ChatGPT. Let's look at the details of GPT-4o mini. GPT-4o mini is based on the same training data and tokens as the flagship GPT-4o model. It has superior reasoning skills and supports up to 128K tokens. The model can process text, images, videos, and audio clips. It can also call multiple functions simultaneously, enabling more efficient data processing with long-context conversations. OpenAI's GPT-4o has higher analytical abilities than Gemini Flash and Claude Haiku. Testing and trials reveal that the AI model generates higher scores in HumanEval and MGSM tests. It reflects that GPT-4o mini has better mathematical reasoning skills, making it faster for coding-related tasks. In addition to performance improvements, GPT-4o mini uses the same safety techniques as GPT-4o to avoid misinformation in its AI-generated responses. It is also the first LLM by OpenAI to feature an instruction hierarchy, that prevents prompt injections, hijacks, and jailbreaks. This improves the AI model's reliability and reduces the chances of potential misuse. GPT-4o mini will be implemented in ChatGPT for all users. Free users will be slowly shifted from GPT-3.5 to GPT-4o mini. The new AI model will also be available as an API for most OpenAI services. The biggest advantage of the GPT-4o mini is for software developers, as this model is up to 60% cheaper than the GPT-3.5 Turbo. Developers will soon be able to use this new model in services like AI support agent bots, which will improve the quality of responses provided by those bots.
[19]
What is GPT-4o Mini, OpenAI's newest launch?
Even a baby spider is cuter than its adult counterpart, so when OpenAI debuted a mini version of its most powerful model, you'll have to excuse my instinct to say "Awwww!" That's probably not always the right response, instincts be damned. Because there isn't very much about this new mini version of GPT-4o that we'd describe as "cute." OpenAI's new AI model, GPT-4o mini, is the AI startup's latest effort to make its chatbot the biggest in the game -- even if it's tiny. The company says it is its "most cost-efficient small model." "We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable," OpenAI said in a press statement. OpenAI went on to specify that the tool currently only supports "text and vision in the API," but said that "support for text, image, video and audio inputs and outputs" will be integrated in the future. It did not specify when that integration will take place. "I think GPT-4o Mini really gets at the OpenAI mission of making AI more broadly accessible to people. If we want AI to benefit every corner of the world, every industry, every application, we have to make AI much more affordable," Olivier Godement, who leads the API platform product, told The Verge. On Thursday, most ChatGPT users -- those who use it on a free, plus, or team plan -- will be able to use GPT-4o Mini instead of GPT-3.5 Turbo. Enterprise users will gain access to the mini model next week.
[20]
Microsoft-backed OpenAI launches cost-efficient GPT 4o-mini AI model - Times of India
Sam Altman-led OpenAI has launched a new cost-efficient small AI model. Dubbed GPT-4o mini, it is 60% cheaper than GPT-3.5 Turbo. GPT 4o-mini is priced at 15 cents per million input tokens and 60 cents per million output tokens. GPT 4o-mini features In a press release, the Microsoft-backed company said that the new GPT-4o mini can outperform GPT-41 on chat preferences in LMSYS leaderboard.It scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. OpenAI's GPT-4o mini enables a broad range of tasks, such as applications that chain or parallelize multiple model calls (e.g., calling multiple APIs), pass a large volume of context to the model (e.g., full code base or conversation history), or interact with customers through fast, real-time text responses (e.g., customer support chatbots). Currently, GPT-4o mini supports text and vision in the API. The company plans to add support for text, image, video and audio inputs and outputs in the future. According to OpenAI, GPT 4o-mini scored 87.0%, compared to 75.5% for Gemini Flash and 71.7% for Claude Haiku in math reasoning. Similarly, GPT-4o mini scored 87.2% on HumanEval, which measures coding performance, compared to 71.5% for Gemini Flash and 75.9% for Claude Haiku. Availability and pricing GPT-4o mini is available as a text and vision model in the Assistants API, Chat Completions API, and Batch API. Developers pay 15 cents per 1M input tokens and 60 cents per 1M output tokens (roughly the equivalent of 2500 pages in a standard book). In ChatGPT, Free, Plus and Team users will be able to access GPT-4o mini starting today, in place of GPT-3.5. Enterprise users will also have access starting next week. Safety measures in GPT-4o mini OpenAI says that the GPT-4o mini has the same safety mitigations built-in as GPT-4o. More than 70 external experts in fields like social psychology and misinformation tested GPT-4o to identify potential risks. Insights from these expert evaluations, the company says, have helped improve the safety of both GPT-4o and GPT-4o mini. The TOI Tech Desk is a dedicated team of journalists committed to delivering the latest and most relevant news from the world of technology to readers of The Times of India. TOI Tech Desk's news coverage spans a wide spectrum across gadget launches, gadget reviews, trends, in-depth analysis, exclusive reports and breaking stories that impact technology and the digital universe. Be it how-tos or the latest happenings in AI, cybersecurity, personal gadgets, platforms like WhatsApp, Instagram, Facebook and more; TOI Tech Desk brings the news with accuracy and authenticity.
[21]
OpenAI brings GPT-4o mini AI model targeting app developers: Check details
OpenAI, on July 18, launched its new small GPT-4o mini AI model, branding it as their most cost-efficient offering. Targeting app developers with its new model, OpenAI in its press note stated that it expects that the GPT-4o mini will expand the range of applications built with AI, as the new model makes AI capabilities more affordable. The Microsoft-backed AI startup company said that the GPT-4o mini surpasses GPT-3.5 Turbo model and other small models from other companies on benchmark scores across textual and multimodal reasoning. Additionally, it stated that the GPT-4o mini enables a variety of tasks with low latency such as passing a large volume of context to the model, interacting with customers with real-time text responses and more.
[22]
OpenAI unveils 'GPT-4o mini' cost-effective AI Model
OpenAI has announced GPT-4o mini (O for Omni), a compact yet powerful AI model aimed at making artificial intelligence more affordable and accessible. It is priced at 15 cents per million input tokens and 60 cents per million output tokens, significantly cheaper than previous models like GPT-3.5 Turbo. OpenAI collaborated with companies like Ramp and Superhuman to explore the model's capabilities. These partners found GPT-4o mini significantly better than GPT-3.5 Turbo for tasks like extracting structured data from receipts or generating high-quality email responses from thread history. OpenAI ensures safety is built into their models from the start, using methods like filtering out unwanted information during pre-training and reinforcement learning with human feedback (RLHF) during post-training. GPT-4o mini has the same safety features as GPT-4o, tested by over 70 external experts in various fields. New techniques like the instruction hierarchy method improve the model's resistance to jailbreaks, prompt injections, and system prompt extractions. OpenAI aims to continue reducing costs while enhancing model capabilities. The cost per token of GPT-4o mini has dropped by 99% since the introduction of text-davinci-003 in 2022. They envision AI models becoming seamlessly integrated into every app and website, making AI more accessible and embedded in daily digital experiences.
[23]
OpenAI launches small AI model GPT-4o Mini. What is it and why is it important?
GPT-4o Mini (O stands for Omni), will replace GPT-3.5 Turbo and will be available to use starting today for free along with ChatGPT Plus and Team members. Meanwhile, it will be offered to enterprise users starting next week. Also Read | OpenAI said that GPT-4o Mini is priced at 15 cents per million input token and 60 cents per million output tokens, making it 60% more cheaper than GPT-3.5 Turbo. The model scored 82% on Massive Multitask Language Understanding (MMLU) and outperformed GPT-4 on chat preferences in LMSYS leaderboard. The company also claimed that GPT-4o Mini also comprehensively defeated other small models in reasoning tasks with Gemini Flash only managing a MMLU score of 77.9% and Claude Haiku a score of 73.8%. Announcing the new model in a blog post, OpenAI wrote, "OpenAI is committed to making intelligence as broadly accessible as possible. Today, we're announcing GPT-4o mini, our most cost-efficient small model. We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable." Small models like GPT-4o require low computational power and hence are a more affordable option for devleopers with limited resources who want to use generative AI in their applications. GPT-4o Mini currently support text and vision in application programming interface (API) and support for text, image, video and audio outputs will be made available in the future, OpenAI said. The latest model has a context window of 128K token, which translates to around 95,000 words, and has a cut off date of October 2023. Meanwhile, OpenAI stated that GPT-4o Mini is even more cost effective handling non-English text now owing to the improved tokenizer. 3.6 Crore Indians visited in a single day choosing us as India's undisputed platform for General Election Results. Explore the latest updates here!
[24]
OpenAI's GPT-4o mini Offers AI Smarts at a Very Low Price
Tech Deals: Anker Nano Power Bank, Pixel Watch 2, Blink Outdoor Cameras, and More! OpenAI is launching a lightweight version of its flagship GPT-4o model. The new miniaturized model, aptly named "GPT-4o mini," will be available for free and paid ChatGPT users later today. Most AI companies, including Google and Meta, offer some kind of lightweight LLM. These tiny models are intended for simple or repetitive tasks. They don't require a ton of processing power, they're relatively inexpensive, and they're pretty fast. So, they're often the best choice for app or software integration. GPT-4o is the most affordable and lightweight model in OpenAI's arsenal. In fact, it's "60% cheaper" than GPT-3.5 Turbo, and it outscores both Gemini 1.5 Flash and Claude 3 Haiku on the MMLU AI benchmark. Exact benchmark scores should be published on Artificial Analysis within the coming days. For those wondering, the "60% cheaper" claim is totally accurate. Developers utilizing GPT-3.5 Turbo currently pay $0.50 for a million input tokens and $1.50 for a million output tokens. Transiting to GPT-4o mini brings per-million token pricing down to $0.15 and $0.60, respectively. That's important because apps that currently offer AI functionality may improve, and developers that previously avoided AI due to cost concerns may be more inclined to join the bandwagon. The new model might also serve as the replacement for GPT-3.5 for ChatGPT users and developers, but that has not yet been confirmed. Average ChatGPT users may see no reason to test GPT-4o mini, which is understandable, since it's currently aimed at developers. However, free and paid ChatGPT users may still get some use out of GPT-4o mini, especially when generating images, proofreading several pages of text, or asking simple questions. A free user may choose the lightweight model to avoid hitting their daily GPT-4o usage cutoff, for example. GPT-4o mini launches this afternoon (July 18th) and will be immediately available to all ChatGPT users. Developers also gain immediate access, though enterprise customers must wait an extra week. Source: OpenAI via TechCrunch
[25]
OpenAI Slashes the Cost of Using Its AI With a "Mini" Model
With competing models -- including many free ones -- flooding the market, OpenAI is announcing a cheaper way to use its AI. OpenAI today announced a cut-price "mini" model that it says will allow more companies and programs to tap into its artificial intelligence. The new model, called GPT-4o mini, is 60 percent cheaper than OpenAI's most inexpensive existing model while offering higher performance, the company says. OpenAI characterizes the move as part of an effort to make AI "as broadly accessible as possible," but it also reflects growing competition among AI cloud providers as well as rising interest in small and free open source AI models. Meta, the social media giant, is expected to debut the largest version of its very capable free offering, Llama 3, next week. "The whole point of OpenAI is to build and distribute AI safely and make it broadly accessible," Olivier Godement, a product manager at OpenAI responsible for the new model tells WIRED. "Making intelligence available at a lower cost is one of the most efficient ways for us to do that." Godement says the company developed a cheaper offering by improving the model architecture and refining the training data and the training regimen. GPT-4o mini outperforms other "small" models on the market in several common benchmarks, OpenAI says. OpenAI has gained a significant foothold in the cloud AI market thanks to interest in tapping into the remarkable capabilities of its chatbot, ChatGPT, which debuted in late 2022. The company lets outsiders access the large language model that powers ChatGPT, called GPT-4o, for a fee. It also offers a less powerful model called GPT-3.5 Turbo for about a tenth of the cost of GPT-4o. The interest in language models triggered by ChatGPT's wild success has prompted competitors to develop similar offerings. Google, a pioneer in AI, has made a major push to build and commercialize a large language model and chatbot under the brand name Gemini. Startups such as Anthropic, Cohere, and AI21 have raised millions to develop and market their own large language models to business customers and developers. Building the highest-performing large language models requires huge financial resources, but some companies have chosen to open source their creations in order to attract developers to their ecosystems. The most prominent open source AI model is Meta's Llama; it can be downloaded and used for free, but its license imposes certain limits on commercial usage. This April, Meta announced Llama 3, its most powerful free model. The company released a small version of the model with eight billion parameters -- a rough measure of a model's portability and complexity -- as well as a more powerful medium-sized, 70 billion parameter version. The medium-sized model is close to OpenAI's best offering on several benchmark scores. Several sources confirmed to WIRED that Meta plans to release the largest version of Llama 3, with 400 billion parameters, on July 23, although they say the release date could change. It is unclear how capable this version of Llama 3 will be, but some companies have turned their attention toward open source AI models because they are cheaper, can be customized, and offer greater control over a model and the data it is fed. Godement concedes that customers' needs are evolving. "What we see more and more from the market is developers and businesses combining small and large models to build the best product experience at the price and the latency that makes sense for them," he says. Godement says OpenAI's cloud offerings provide customers with models that have gone through more security testing than competitors'. He adds that OpenAI could eventually develop models that customers can run on their own devices. "If we see massive demand, we may open that door," he says.
[26]
OpenAI unveils GPT-4o mini -- a smaller, much cheaper multimodal AI model
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More A little more than two months ago, OpenAI released GPT-4o, its newest and most powerful AI model that was the first to be trained by the company natively to handle multimodal inputs and outputs (text, image, audio, and ultimately video) without linking to other models for help. It was the most powerful, publicly available AI model in the world on third-party benchmarks upon release, but was outclassed shortly after by rival Anthropic's Claude 3.5 Sonnet a few weeks later, and the two have been neck-and-neck ever since. But OpenAI isn't stopping there: today, it is announcing a smaller version of that model, GPT-4o mini, which it says is "the most cost-efficient small model in the market," costing developers just $0.15 USD per 1 million tokens a user inputs, and $0.60 for every million they receive back from the model, for third-party apps and services built atop it using OpenAI's application programming interfaces (APIs). It's also far cheaper than GPT-4o, which costs $5.00 for 1 million input tokens and $15 per 1 million output tokens. Tokens, as you'll recall, are the numerical codes that represent semantic units, words, numbers, and other data inside a given large language model (LLM) or small language model (SML) -- the latter which mini appears to be (OpenAI did not release the number of parameters, or connections between artificial neurons, the model has, making it difficult to say how large or small it is, but the mini name clearly gives an indication.) Olivier Godement, OpenAI's Head of Product, API, told VentureBeat in a teleconference interview yesterday that GPT-4o mini is particularly helpful for enterprises, startups and developers "building any agent" from "a customer support agent" to "a financial agent," as those typically perform "many calls back to the API," resulting in a high volume of tokens inputted and outputted by the underlying source model, which can quickly drive up costs. "The cost per intelligence is so good, I expect it's going to be used for all sorts of customer support, software engineering, creative writing, all kinds of tasks," said Godement. "Every time we adopt a new model, there are new cases that pop up, and I think that will be even more the case for GPT-4o mini." The move to launch GPT-4o mini also comes ahead of Meta's reported release of its massive Llama 3 400-billion parameter model expected next week, and seems quite clearly designed to pre-empt that news and cement in developers' minds that OpenAI remains the leader in enterprise-grade AI. 60% cheaper than GPT-3.5 Turbo for developers To put GPT-4o mini's cost into perspective, it's 60% less than GPT-3.5 Turbo, previously the most affordable model among OpenAI's offerings since the release of GPT-4o. At the same time, the model is targeted to be as fast at working as GPT-3.5 Turbo, transmitting around 67 tokens per second. OpenAI is pitching GPT-4o mini as a direct successor to GPT-3.5 Turbo, but a much more capable one, as it also can handle text and vision inputs, unlike GPT-3.5 Turbo, which could only handle text. At some point in the future, OpenAI says GPT-4o mini will also be able to generate imagery and other multimodal outputs including audio and video, as well as accept them as inputs. But for now, only the text and still image/document inputs will be available today. At present, GPT-4o mini outperforms GPT-3.5 Turbo on a range of third-party benchmarks, other comparably classed models such as Google's Gemini 1.5 Flash and Anthropic's Claude 3 Haiku, and even GPT-4 itself on some tasks. Specifically, OpenAI released benchmarks showing that GPT-4o mini scores 82.0% on the Massive Multitask Language Understanding (MMLU) benchmark, which includes multiple choice questions about subjects from math, science, history, and more, versus 77.9% for Gemini Flash and 73.8% for Claude Haiku. Coming to Apple devices this fall as well In addition, Godement told VentureBeat that GPT-4o mini would be available this fall through Apple Intelligence, the new AI service from Apple Inc., for its mobile devices and Mac desktops, timed to coincide with the release of its new iOS 18 software, as part of the partnership between OpenAI and Apple announced at the latter's WWDC event last month. However, the model will still be running on OpenAI cloud servers -- not on device, which would seem to negate one of the advantages of running a small model in the first place, a local inference that is by nature, faster, more secure, and doesn't require a web connection. Yet Godement pointed out that even when connecting to OpenAI cloud servers, the GPT-4o mini model is faster than others available from the company. Moreover, he told VentureBeat that most third-party developers OpenAI worked with were not yet interested in running the company's models locally, as it would require much more intensive setup and computing hardware on their end. However, the introduction of GPT-4o mini raises the possibility that OpenAI developer customers may now be able to run the model locally more cost effectively and with less hardware, so Godement said it was not out of the question that such a solution could one day be provided. Replacing GPT-3.5 turbo in ChatGPT, but not killing it entirely for developers Beginning later today, GPT-4o mini will replace GPT-3.5 Turbo among the options for paying subscribers of ChatGPT including the Plus and Teams plans -- with support for ChatGPT Enterprise coming next week. The model will appear in the drop-down menu on the upper left corner of the web and Mac desktop apps. However, ChatGPT users won't get a price reduction on their paid subscriptions for selecting GPT-4o mini -- only developers building atop the API will benefit from the savings. Yet ChatGPT users will have access to a newer, faster, and more powerful model for tasks than GPT-3.5 Turbo automatically, which is certainly a benefit. OpenAI isn't yet deprecating or phasing out support for GPT-3.5 Turbo in its APIs, as the company doesn't want to force developers to upgrade or to break the apps that are currently built atop this older model. Instead, the company believes that developers will likely naturally migrate quickly en masse to using the new model since it is a significant cost reduction and boost in intelligence and other capabilities. Some developers have already been alpha testing GPT-4o mini, according to Godement, including enterprise expense management and accounts software startup Ramp and the cloud email AI startup Superhuman, and both are said to have reported excellent results. Godement said GPT-4o mini is powering Ramp's automatic receipt categorization and merchant detection features, and powering Superhuman's suggested, custom-tailored email responses. Ramp in particular has "seen pretty amazing results for its data extraction tests," from receipts, said Godement. He was not able to say precisely whether Ramp was using the GPT-4o mini native multimodal vision input or if the firm was using another system to first extract text and numerals from receipts and send it to the model. So why should any developers still use the older, more expensive GPT-4o parent model? Given the significant cost savings offered by GPT-4o mini and high performance benchmarks on a number of tasks and tests, the question naturally arises: why would a developer pay more money to use the full GPT-4o model when the mini one is now available? OpenAI believes that for the most computationally-intensive, complex, and demanding applications, the full GPT-4o is still the way to go, and justifies its higher price in comparison. "Let's assume I'm building medical applications that I'd like to summarize and propose some diagnosis for patients," Godement gave as one example. "I'm basically going to optimize for intelligence. I want to make sure they get the most intelligent model out of the box. Similarly, if you're building a software engineering assistant and working on a pretty complex codebase, you will still be see better results with GPT-4o. If intelligence differentiates your product, I recommend you stick with GPT-4o and you'll get the best results."
[27]
OpenAI's GPT-4o mini model is battling Gemini Flash and Claude Haiku
The GPT-4o mini (its mini, not Mini) is keeping the size and weight in check, as is often the case with AI models and how they're used, for broader relevance. Even then, OpenAI insists the cost-effective small model is better than even GPT-4 in many tasks, whilst being significantly less costly as well. Competitors too, on the MMLU and MMMU benchmarks. For instance, in the MMLU benchmark that tests multimodal reasoning, GPT-4o mini scores an 82% accuracy rate, while Gemini Flash (77.9%) and Claude Haiku (73.9%) are far behind. "Today, GPT-4o mini supports text and vision in the API, with support for text, image, video and audio inputs and outputs coming in the future. The model has a context window of 128K tokens, supports up to 16K output tokens per request, and has knowledge up to October 2023. Thanks to the improved tokenizer shared with GPT-4o, handling non-English text is now even more cost effective," the company says in a statement shared with HT. The pricing they're indicating for developers is 15 cents per million input tokens and 60 cents per million output tokens - this is, according to official numbers, 60% less costly than the GPT-3.5 Turbo. In due course, GPT-4o mini in the API will have full multimodal support like GPT-4o, supporting text, image, video, and audio inputs and outputs. OpenAI had to tackle the small model class effectively, and largely two reasons. First was the pricing, to stay competitive. The way things are brewing for developers requiring a more cost-effective AI model to integrate within their apps or platforms, they'd have an option to choose from Google Gemini 1.5 Flash or Anthropic's Claude 3 Haiku. The Claude 3 Haiku, for example, is priced at 25 cents per million tokens input and 1.25 cents per million tokens output. Secondly, the need for different sized models, for flexibility with size, capabilities, performance, utility and efficiency. Google has the Gemini in Ultra, Pro and Flash model sizes. Anthropic has the Claude 3.5 in Sonnet, Opus and Haiku sizes. Gemini Nano's on-device specifics may not be relevant to tackle in OpenAI's case, at least till the iOS, iPadOS and macOS integration comes through. For now, GPT-4o mini is available as a text and vision model in the Assistants API, Chat Completions API, and Batch API for developers to integrate. OpenAI confirms to HT that the plan is to roll out fine-tuning for GPT-4o mini in the coming days. For everyone else, access to GPT-4o mini is now available for Free, Plus and Team users, while Enterprise subscribers get access in the coming next week. For OpenAI, and indeed the entire AI space, this hasn't been a summer to forget. The subplots played their part in the intrigue. Google talked about an updated Gemini 1.5 Pro model, along with updates to Gemini Nano, and Gemini Live (in early May) which will allow AI and you to see the world through a phone's camera. OpenAI had an advance response to everything Google announced, by unleashing the new GPT-4o model, as well as real-world views for context. An example of that being OpenAI's partnership with Be My Eyes, which has upgraded from GPT-4 to GPT-4o as the foundation for guidance for users with visual impairments. Then came the big announcement with Apple. The fact that Google responded to OpenAI's GPT-4o's ability to see the world through the phone camera with Gemini Live, within 24 hours, details just how little margin there is to get any of this wrong. And how competitive the space is. None of that is likely to change, anytime soon.
[28]
OpenAI's GPT-4o Mini isn't much better than rival LLMs
Plus: Meta Euro model drama, Mistral and Nvidia find NeMo, and more AI Roundup OpenAI has made available GPT-4o Mini, a smaller and cheaper version of its GPT-4o generative large language model (LLM), via its cloud. The Microsoft-backed super lab said Thursday GPT-4o Mini is like regular GPT-4o in that it's multimodal - it can handle more than just the written word - and has a context window of 128,000 tokens and was trained on materials dated up to October 2023. The Mini can emit up to 16,000 tokens of output. While GPT-4o, OpenAI's top-end model, costs $5 and $15 per million input and output tokens, respectively, the Mini edition costs 15 and 60 cents, again respectively. You can halve those numbers if using delayed batch processing. We're told the cut-down version is not fully featured yet, supporting just text and vision via its API, with other input and output formats, such as audio, coming in the indeterminate future. In creating GPT-4o Mini, OpenAI emphasized how safe it had made the thing, claiming to filter out offensive data from training materials and giving it the same guardrails that GPT-4o has. Furthermore, OpenAI claimed GPT-4o Mini is ahead of comparable LLMs in benchmarks. Indeed, compared to Google's lighter-weight Gemini Flash and Anthropic's Claude Haiku, Mini was usually between five and 15 percent more accurate in tests such as MMLU, though in two outliers it was nearly twice as accurate as the competition and in another a little worse than Gemini Flash but still ahead of Claude Haiku, allegedly. It's particularly personal for OpenAI to seemingly beat Anthropic as the latter was co-founded and built by executives and engineers from the former plus others. That GPT-4o Mini looks good in the graph above for sure, though it doesn't have an overall commanding lead, and that's indicative of OpenAI's recent loss of absolute leadership in the modern LLM arena. As veteran open source developer Simon Willison detailed in his keynote at the AI Engineer World's Fair last month, 2024 has seen many of OpenAI's competitors release their own GPT-4o-class models. "The best models are grouped together: GPT-4o, the brand new Claude 3.5 Sonnet and Google Gemini 1.5 Pro," Willison said. "I would classify all of these as GPT-4 class. These are the best available models, and we have options other than GPT-4 now. The pricing isn't too bad either -- significantly cheaper than in the past." At 82 percent accuracy in MMLU and a cost of 15 cents per million tokens, GPT-4o Mini is mostly ahead of the pack. However, Willison says the LMSYS Chatbot Arena benchmark provides a more realistic evaluation of LLM quality because actual humans are asked to compare outputs and choose which is better, a brute-force but effective way of ranking different models. GPT-4o Mini is too new to be included in the tournament-style benchmark, though he notes that full-size GPT-4o is only barely ahead of its rivals. Anthropic's flagship Claude 3.5 Sonnet currently has 1,271 points to GPT-4o's 1,287. Gemini 1.5 Pro isn't far behind at 1,267; slightly less performant but still respectable models include Nvidia and Mistral's brand-new Nemotron 4 340B Instruct at 1,209 points, and Meta's LlaMa 3 70B Instruct at 1,201. Willison also noted the Mini is cheaper than Claude 3 Haiku and Gemini 1.5 Flash. OpenAI may be the best, in terms of these test scores, from small to big LLMs, though it no longer has the dominating lead it once had. That's probably a good thing; between costly AI hardware and high power usage, the last thing AI needed was a monopoly on LLMs. ®
Share
Share
Copy Link
OpenAI introduces GPT-4o Mini, a smaller and more affordable version of its AI language model. This new offering aims to make advanced AI technology more accessible to individuals and businesses with limited resources.
OpenAI, the artificial intelligence research laboratory, has announced the release of GPT-4o Mini, a smaller and more cost-effective version of its renowned language model. This new AI chatbot is designed to be faster and more accessible, potentially revolutionizing the AI landscape for individuals and businesses alike 1.
GPT-4o Mini boasts several improvements over its predecessors. The model is significantly smaller, which translates to faster processing times and reduced computational requirements. This efficiency allows for quicker responses and lower operational costs, making it an attractive option for a wider range of users 2.
One of the most notable aspects of GPT-4o Mini is its affordability. OpenAI has priced this model at a fraction of the cost of its larger counterparts, making advanced AI technology more accessible to individuals and smaller organizations with limited budgets 3.
The introduction of GPT-4o Mini is expected to have far-reaching implications across various sectors. From enhancing customer service chatbots to powering more efficient content creation tools, the model's versatility opens up new possibilities for AI integration in diverse fields 4.
Experts predict that the availability of a more affordable AI model could lead to increased adoption among small and medium-sized enterprises, potentially leveling the playing field in industries where AI capabilities have become a competitive advantage 5.
While GPT-4o Mini presents numerous benefits, it's important to note that the reduced size may come with some limitations in terms of knowledge breadth and task complexity compared to larger models. OpenAI has stated that they have optimized the model to maintain high performance on common tasks while significantly reducing its size 1.
The announcement of GPT-4o Mini has generated considerable buzz in the tech industry. Competitors are likely to follow suit with their own smaller, more accessible AI models, potentially sparking a new trend in AI development focused on efficiency and affordability 3.
As AI technology continues to evolve, the introduction of models like GPT-4o Mini represents a significant step towards democratizing access to advanced AI capabilities. This development could accelerate innovation across various sectors and pave the way for more inclusive AI-driven solutions in the future 5.
Reference
[1]
[2]
[5]
OpenAI has introduced GPT-4o Mini, a lightweight version of its flagship AI model. This new model promises improved speed and lower costs while maintaining high performance, potentially revolutionizing AI accessibility.
8 Sources
OpenAI has launched GPT-4O Mini, a more affordable version of its top AI model. This new offering aims to make advanced AI technology more accessible to developers and businesses while potentially reshaping the competitive landscape in the AI industry.
5 Sources
OpenAI introduces GPT-4O Mini, a compact version of its large language model. This new AI aims to be more efficient and accessible while maintaining impressive capabilities.
2 Sources