Curated by THEOUTPOST
On Fri, 19 Jul, 12:02 AM UTC
24 Sources
[1]
Goodbye GPT-3.5, OpenAI's new GPT-4o mini AI model is all about compact power
OpenAI has added a new large language model (LLM) called GPT-4o mini to ChatGPT and its APIs. As the name implies, the GPT-4o Mini model is a smaller version of the GPT-4o model introduced in May. The mini model is designed to balance the power of GPT-4o with a more cost-efficient approach. GPT-4o mini has much of the functionality of its larger cousin, though the API only has text and vision support for now, with image, video, and audio inputs and outputs still in the works. Like GPT-4o, the new model has a context window of 128,000 tokens, or eight times that of GPT-3.5 Turbo. The new model also comes with enhanced safety features. Along with those built into GPT-4o already, GPT-4o mini added new techniques that make it more resistant to jailbreaks and improper prompt injections, among other issues concerning developers looking to deploy AI APIs broadly. OpenAI suggests the bigger context window and other upgrades, such as improved non-English text understanding, will make GPT-4o mini especially useful for processing big documents or linking multiple interactions with the AI model. For example, it could provide better recommendations in online stores, speed up real-time text responses for customer service, and produce accurate and detailed answers to students studying for an exam more quickly than other models. OpenAI has visions of GPT-4o automating and streamlining business processes thanks to its ability to fetch data and take actions with external systems. For businesses using the API, the cost is notably reduced to just over half the price per token of GPT-3.5 Turbo. "OpenAI is committed to making intelligence as broadly accessible as possible," OpenAI explained in its announcement. "We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable." GPT-4o mini is part of the recent wave of smaller LLMs like Google's Gemini Flash and Anthropic's Claude Haiku. According to OpenAI, however, GPT-4o mini blows them out of the water when it comes to many of the standard tests. The model scored 82% on the Massive Multitask Language Understanding (MMLU) benchmark, compared to 77.9% and 73.8% by Gemini Flash and Haiku, respectively. The same goes for the MGSM and Human Eval tests, where GPT-4o Mini hit 87% and 87.2%, while Gemini Flash had 75.5% and 71.5%, and Haiku had 71.7% and 75.9%. In other words, GPT-4o Mini wins out on textual comprehension in addition to math and coding tasks, as can be seen in the graph below. The introduction of GPT-4o Mini represents a significant step in making advanced AI more affordable and accessible, according to OpenAI. Lower costs plus better performance will likely help incorporate AI into everyday applications. The same goes for ChatGPT users, who can all access the model starting this week. OpenAI also has plans to introduce fine-tuning capabilities for GPT-4o Mini within the API. The broader picture shows another step in ChatGPT's evolving services. As OpenAI phases out GPT-3.5 for ChatGPT, the focus shifts to the next stage of providing more powerful models. OpenAI CEO Sam Altman has long hinted at how GPT-5 will "substantially improve" upon the existing models. At the same time, the leaked OpenAI scale for measuring AI power shows there is still a long way to go to the still-mythical artificial general intelligence (AGI) that can perfectly mimic the workings of the human mind.
[2]
OpenAI launches GPT-4o mini- cheap yet powerful version of GPT-4
OpenAI is making AI more accessible with GPT-4o mini, a new cost-effective small language model. This new model boasts superior performance in text and code compared to previous options, all at a reduced price. Previously, the high cost of AI models acted as a barrier for many developers. GPT-4o mini breaks this barrier with a price tag of just 15 cents per million input tokens and 60 cents per million output tokens. Benchmarks reveal that the GPT-4o mini outperforms other small models in crucial areas like textual reasoning, mathematical problem-solving, and even coding proficiency. This versatility extends to multimodal tasks, where the model can handle scenarios involving both text and images. The language model is trained to meticulously filtered to remove harmful content, and the model is further fine-tuned using reinforcement learning to ensure accurate and reliable responses. Additionally, a novel instruction hierarchy method safeguards against misuse, making GPT-4o mini a trustworthy tool for developers. Currently, GPT-4o mini is available through OpenAI's APIs for text and vision tasks. To further empower developers, OpenAI plans to introduce fine-tuning capabilities in the near future, allowing for even greater control over the model's outputs.
[3]
OpenAI offers GPT-4o mini to slash the cost of applications
OpenAI on Thursday unveiled a stripped-down version of its GPT-4o large language model, GPT-4o mini, which it said has better accuracy than GPT-4 on tasks, and costs dramatically less than GPT-3.5 "Turbo" when used by developers, which it said can boost the construction of applications that use the AI model extensively. The company touts the new AI model as "the most cost-efficient small model in the market," although, as with most OpenAI releases, no technical details are available about GPT-4o mini, such as the number of parameters, hence, it's unclear what "small" means in this case. (An "AI model" is the part of an AI program that contains numerous neural net parameters and activation functions that are the key elements for how an AI program functions.) Also: How to use ChatGPT to create an app GPT-4o mini "is priced at 15 cents per million input tokens and 60 cents per million output tokens, an order of magnitude more affordable than previous frontier models and more than 60% cheaper than GPT-3.5 Turbo," said OpenAI in a blog post emailed to ZDNET. That reduction in cost, said the company, will aid the development of applications that are affected by volume of activity. For example, applications that must make multiple API (application programming interface) calls, or that use larger "context windows" to retrieve materials (say, to retrieve an entire code-base when developing an app), or that have to interact frequently with the end user, such as a help desk support bot, will benefit from the reduction in per-transaction cost, said OpenAI. The model, says OpenAI, outperforms the standard GPT-4 model when used as a chatbot, based on crowd-sourced tests by the Lmsys leaderboard. It also "surpasses GPT-3.5 Turbo and other small models on academic benchmarks across both textual intelligence and multimodal reasoning," and supports as many languages as the standard GPT-4o model. The new model is available immediately to developers via the Assistants API, Chat Completions API, and Batch API, and can be used instead of GPT-3.5 Turbo in ChatGPT's free, plus, and team accounts. The model offers only text and image support at the moment, with audio and video to be added at an unspecified date. The GPT-4o mini context window is 128,000 tokens, and its training data is current through October of 2023.
[4]
OpenAI Unveils Cheaper Small AI Model GPT-4o Mini
(Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said. (Reporting by Deborah Sophia in Bengaluru; Editing by Vijay Kishore)
[5]
OpenAI unveils cheaper small AI model GPT-4o mini
July 18 (Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said. (Reporting by Deborah Sophia in Bengaluru; Editing by Vijay Kishore)
[6]
OpenAI released a smaller, cheaper version of its most powerful AI model
OpenAI introduced a smaller, more cost-efficient version of its most powerful model GPT-4o -- and it could mean even more AI-powered apps. The new model, GPT-4o mini, is smarter and more than 60% cheaper than GPT-3.5 Turbo, according to OpenAI, meaning it is more affordable for developers building applications on OpenAI's models. GPT-4o mini is priced at 15 cents per 1 million input tokens -- or the unit of information an AI model uses, for example, a word or phrase -- and 60 cents per 1 million output tokens. That is equivalent to 2,500 pages in a book, OpenAI said. "We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable," OpenAI said in a statement. On the Massive Multitask Language Understanding (MMLU) benchmark, which measures language models' capabilities across 57 subjects including math and law, GPT-4o mini scored 82%, outperforming other small models and GPT-3.5 Turbo, according to OpenAI. The small model will replace GPT-3.5 Turbo in ChatGPT, and Free, Plus, and Team users have access to the new model now. At its release, GPT-4o mini can support text and vision in the API, or application programming interface, which allows developers to build their own AI models off of OpenAI's technology. In the future, GPT-4o will have support for text, image, video, and audio, OpenAI said. The small model has knowledge up to October 2023. "We envision a future where models become seamlessly integrated in every app and on every website," OpenAI said. "GPT-4o mini is paving the way for developers to build and scale powerful AI applications more efficiently and affordably." OpenAI unveiled ChatGPT-4o, in May, which is multimodal, meaning it can see, hear, and have real-time conversations.
[7]
OpenAI Previews Slimmed-Down and Cheaper GPT-4o Mini Model
OpenAI is launching GPT-4o Mini, a scaled-down and cheaper version of newest GPT-4o model. GPT-4o Mini can accept and output text and image content; over time, it will be able to process other types of requests, Bloomberg reports. It replaces GPT-3.5 Turbo in ChatGPT and is available to those using the free and paid versions of OpenAI's chatbot. Enterprise users get it next week. For developers, GPT-4o Mini will be a more affordable way to use OpenAI's tech. They'll pay $0.15 per million input tokens and $0.60 per million output tokens -- 60% cheaper than GPT-3.5 Turbo, TechCrunch reports. It has a context window of 128,000 tokens and is trained on data up to October 2023. At I/O in May, Google also launched a smaller AI model "optimized for narrower or high-frequency tasks," dubbed Gemini 1.5 Flash. At OpenAI, the company is aiming for a better multi-modal experience. "If you think about the way we as humans process the world and engage with the world, we see things, we hear things, we say things -- the world is much bigger than text," OpenAI COO Brad Lightcap tells CNBC. OpenAI launched GPT-4o earlier this year with promises of it being the most humanlike AI yet thanks to features like Voice Mode, which is currently delayed while OpenAI works out some kinks in the code. Look for it later this year.
[8]
OpenAI unveils cheaper small AI model GPT-4o mini
July 18 (Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy-intensive, allowing the startup to target a broader pool of customers. Microsoft-backed (MSFT.O), opens new tab OpenAI, the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals such as Meta (META.O), opens new tab and Google (GOOGL.O), opens new tab rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. With the mini model currently supporting text and vision in the application programming interface, OpenAI said support for text, image, video and audio inputs and outputs would be made available in the future. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini -- which has knowledge up to October 2023 -- starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said. Reporting by Deborah Sophia in Bengaluru; Editing by Vijay Kishore Our Standards: The Thomson Reuters Trust Principles., opens new tab
[9]
OpenAI unveils cheaper small AI model GPT-4o mini
MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage.ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy-intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals such as Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. With the mini model currently supporting text and vision in the application programming interface, OpenAI said support for text, image, video and audio inputs and outputs would be made available in the future. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini - which has knowledge up to October 2023 - starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said.
[10]
OpenAI Introduces GPT-4o Mini, 30x Cheaper than GPT-4o
Priced at 15 cents per million input tokens and 60 cents per million output tokens, GPT-4o mini is 30x cheaper than GPT-40 and 60% cheaper than GPT-3.5 Turbo. OpenAI chief Sam Altman made a cost comparison, saying, "Way back in 2022, the best model in the world was text-davinci-003. It was much, much worse than this new model. It cost 100x more." The model excels in various tasks, including text and vision, and supports a context window of 128K tokens with up to 16K output tokens per request. GPT-4o mini demonstrates superior performance on benchmarks, scoring 82% on the MMLU, 87% on MGSM for math reasoning, and 87.2% on HumanEval for coding tasks. It outperforms other small models like Gemini Flash and Claude Haiku in reasoning, math, and coding proficiency. GPT-4o mini's low cost and latency enable a wide range of applications, from customer support chatbots to API integrations. It currently supports text and vision, with future updates planned for text, image, video, and audio inputs and outputs. Safety measures are integral to GPT-4o mini, incorporating techniques like reinforcement learning with human feedback (RLHF) and the instruction hierarchy method to improve model reliability and safety. GPT-4o mini is now available in the Assistants API, Chat Completions API, and Batch API. It will be accessible to Free, Plus, and Team users in ChatGPT today, and to Enterprise users next week. Fine-tuning capabilities will be introduced soon. GPT-4o mini comes after OpenAI co-founder Andrej Karpathy recently demonstrated how the cost of training large language models (LIMs) has significantly decreased over the past five years, making it feasible to train models like GPT-2 for approximately $672 on "one 8XH100 GPU node in 24 hours". "Incredibly, the costs have come down dramatically over the past five years due to improvements in compute hardware (H100 GPUs), software (CUDA, cuBLAS, cuDNN, FlashAttention) and data quality (e.g., the FineWeb-Edu dataset)," said Karpathy. That explains how Tech Mahindra was able to build Project Indus for well under $5 million, which again, is built on GPT-2 architecture, starting from the tokeniser to the decoder. It would be interesting to see what innovative applications developers will create using this new AI model. Looks like it's already in motion. A few days back a mysterious model had appeared on the Chatbot Arena. Unsurprisingly, that model is none other than GPT-4o mini.
[11]
OpenAI Announces GPT-4o Mini: Cheaper And Faster AI Chatbot For More People
OpenAI wants more people to pay for ChatGPT and a cheaper model makes sense (Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market. Priced at 15 cents per million input tokens and 60 cents per million output tokens, the GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo, OpenAI said. It currently outperforms the GPT-4 model on chat preferences and scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. MMLU is a textual intelligence and reasoning benchmark used to evaluate the capabilities of language models. A higher MMLU score signifies it can understand and use language better across a variety of domains, enhancing real-world usage. The GPT-4o mini model's score compared with 77.9% for Google's Gemini Flash and 73.8% for Anthropic's Claude Haiku, according to OpenAI. Smaller language models require less computational power to run, making them a more affordable option for companies with limited resources looking to deploy generative AI in their operations. ChatGPT's Free, Plus and Team users will be able to access GPT-4o mini starting Thursday, in place of GPT-3.5 Turbo, with enterprise users gaining access starting next week, OpenAI said.
[12]
OpenAI unveils 'GPT-4o mini' cost-effective AI Model
OpenAI has announced GPT-4o mini (O for Omni), a compact yet powerful AI model aimed at making artificial intelligence more affordable and accessible. It is priced at 15 cents per million input tokens and 60 cents per million output tokens, significantly cheaper than previous models like GPT-3.5 Turbo. OpenAI collaborated with companies like Ramp and Superhuman to explore the model's capabilities. These partners found GPT-4o mini significantly better than GPT-3.5 Turbo for tasks like extracting structured data from receipts or generating high-quality email responses from thread history. OpenAI ensures safety is built into their models from the start, using methods like filtering out unwanted information during pre-training and reinforcement learning with human feedback (RLHF) during post-training. GPT-4o mini has the same safety features as GPT-4o, tested by over 70 external experts in various fields. New techniques like the instruction hierarchy method improve the model's resistance to jailbreaks, prompt injections, and system prompt extractions. OpenAI aims to continue reducing costs while enhancing model capabilities. The cost per token of GPT-4o mini has dropped by 99% since the introduction of text-davinci-003 in 2022. They envision AI models becoming seamlessly integrated into every app and website, making AI more accessible and embedded in daily digital experiences.
[13]
OpenAI's GPT-4o mini launched | Smaller and cheaper than GPT-3.5 Turbo model
OpenAI announced the release of GPT-4o mini, which it called its "most cost-efficient small model." GPT-4o mini can support text and vision in the API, while support for text, image, video and audio inputs and outputs is yet to come. Per the ChatGPT-maker, GPT-4o mini has a context window of 128K tokens and supports up to 16K output tokens per request. Its knowledge cut-off is October 2023 and the model handles non-English text in a more cost-effective way, claimed the company. While the name might be "mini," OpenAI stressed that the small model could hold its own against both smaller rivals as well as provide an experience comparable to larger ones. (For top technology news of the day, subscribe to our tech newsletter Today's Cache) AI accessibility? Blind gamer puts ChatGPT to the test "GPT-4o mini surpasses GPT-3.5 Turbo and other small models on academic benchmarks across both textual intelligence and multimodal reasoning, and supports the same range of languages as GPT-4o," said OpenAI. ChatGPT's Free, Plus, and Team users can access the new model immediately, while Enterprise users will get access from next week. OpenAI noted that safety measures were in place from the pre-training stage so that the model would not learn from hate speech, adult content, sites that primarily aggregate personal information, and spam. In addition, the model has been fortified to better stand against jailbreak attempts, prompt injections, and system prompt extractions. "GPT-4o mini surpasses GPT-3.5 Turbo and other small models on academic benchmarks across both textual intelligence and multimodal reasoning, and supports the same range of languages as GPT-4o. It also demonstrates strong performance in function calling, which can enable developers to build applications that fetch data or take actions with external systems, and improved long-context performance compared to GPT-3.5 Turbo," said OpenAI in its statement introducing the new model. The AI company backed by Microsoft was criticised by whistleblowers and former employees who claimed that it did not take enough safety precautions when releasing new products, and that it tried to stop employees from speaking up about the same. Read Comments
[14]
OpenAI Now Has a GPT-4o Mini. Here's Why That Matters
ChatGPT maker OpenAI introduced a smaller model called GPT-4o Mini on Thursday, which it says is smarter and cheaper than GPT-3.5 Turbo, an earlier model that was built for simple tasks like dialogue. OpenAI hopes developers will use GPT-4o Mini to "significantly expand the range of applications built with AI," according to a blog post. Chatbots like ChatGPT are the interface we use to communicate with large language models, or LLMs, like GPT-4o Mini and the original, much larger GPT-4o. These models are trained to understand how we use language so they can generate content that sounds human. An LLM can have 1 billion or more parameters, which is a measure of how much content it can ingest before producing a response to your prompt. That means LLMs can learn from and understand a lot, but they aren't ideal for every situation. They can be expensive and consume a lot of energy because of the need for expansive server farms and access across the cloud. A small language model is a compromise of sorts. It offers AI horsepower and speed but doesn't require the same computing resources or cost. Microsoft's Phi-3 Mini, which is built to run on phones and PCs, is one example. Google's Gemini 1.5 Flash, which is designed for high-volume, high-frequency tasks like generating captions and extracting data from forms, is another. Now we have GPT-4o Mini as well. Both free and paid ChatGPT users can access GPT-4o Mini starting Thursday in place of GPT-3.5, which was released in November 2022. GPT-4o Mini currently supports text and vision in the OpenAI API, which is what developers use to build new applications based on OpenAI technology. Support for text, image, video and audio inputs and outputs is "coming in the future," the post said. Enterprise users will have access to GPT-4o Mini starting the week of July 22. OpenAI said GPT-4o Mini excels in mathematical reasoning and coding, but has also demonstrated skills in tasks that require reasoning. Financial tech startup Ramp and email app Superhuman tested out GPT-4o Mini to extract data from files and generate email responses, according to the post. The new model has a context window of 128,000 tokens, which is a measurement of how much it can remember in a given conversation. By way of comparison, GPT-4o has the same context window, while GPT-3.5 Turbo has a context window of 16,000 tokens. GPT-4o Mini costs 15 cents per million input tokens and 60 cents per million output tokens, which OpenAI said is about equal to 2,500 pages in a book. GPT-4o, which was released in May, costs $5 per million input tokens and $2.50 per million output tokens. "We envision a future where models become seamlessly integrated in every app and on every website," the blog post said. "GPT-4o mini is paving the way for developers to build and scale powerful AI applications more efficiently and affordably."
[15]
OpenAI Launches New GPT-4o mini AI Model; Faster and More Efficient Than GPT 3.5 Turbo | MySmartPrice
Free users of ChatGPT will now have access to higher and more accurate responses. OpenAI has launched its new AI model GPT-4o mini with improved multimodal reasoning. This new large language model (LLM) is faster and supports longer contextual conversations compared to GPT 3.5 Turbo. It will be available for both free and paid users of ChatGPT. Let's look at the details of GPT-4o mini. GPT-4o mini is based on the same training data and tokens as the flagship GPT-4o model. It has superior reasoning skills and supports up to 128K tokens. The model can process text, images, videos, and audio clips. It can also call multiple functions simultaneously, enabling more efficient data processing with long-context conversations. OpenAI's GPT-4o has higher analytical abilities than Gemini Flash and Claude Haiku. Testing and trials reveal that the AI model generates higher scores in HumanEval and MGSM tests. It reflects that GPT-4o mini has better mathematical reasoning skills, making it faster for coding-related tasks. In addition to performance improvements, GPT-4o mini uses the same safety techniques as GPT-4o to avoid misinformation in its AI-generated responses. It is also the first LLM by OpenAI to feature an instruction hierarchy, that prevents prompt injections, hijacks, and jailbreaks. This improves the AI model's reliability and reduces the chances of potential misuse. GPT-4o mini will be implemented in ChatGPT for all users. Free users will be slowly shifted from GPT-3.5 to GPT-4o mini. The new AI model will also be available as an API for most OpenAI services. The biggest advantage of the GPT-4o mini is for software developers, as this model is up to 60% cheaper than the GPT-3.5 Turbo. Developers will soon be able to use this new model in services like AI support agent bots, which will improve the quality of responses provided by those bots.
[16]
ChatGPT maker OpenAI goes smaller and cheaper with new AI tech
The new version of its AI model, called GPT-4o mini, is 60% cheaper to use than the model that powered ChatGPT until recently, called GPT-3.5 turbo, the company said, and scores higher on tests it uses to measure intelligence and effectiveness. OpenAI is the best-known brand in the booming space of generative artificial intelligence, so its embrace of efficiency is a notable indicator of a broader turn for the industry. Previously, OpenAI and competitors like Google and Microsoft focused primarily on developing the biggest and most powerful AI models. Now they are trying to balance those pushes with releasing small models that can be more profitable for them and more useful for business clients that only need AI tech to do certain tasks well. Google, and startups Anthropic, Mistral and Cohere have also released smaller models this year. Microsoft has played up a family of small models named Phi it said was 1/100th the size of the model behind ChatGPT at the time. Many generative AI startups, as well as new entrant Apple, are also focusing on small models that are cheaper to build because they are typically trained on smaller batches of data. They also require less computing power to operate and can in some cases work directly off hardware like a phone. OpenAI said GPT-4o mini can currently interpret text and image inputs and will later add the capability to scan and output audio and video. The technology was rolled out to people who use the free and premium versions of ChatGPT starting Thursday. The company said it would be available to corporate customers next week.
[17]
Microsoft-backed OpenAI launches cost-efficient GPT 4o-mini AI model - Times of India
Sam Altman-led OpenAI has launched a new cost-efficient small AI model. Dubbed GPT-4o mini, it is 60% cheaper than GPT-3.5 Turbo. GPT 4o-mini is priced at 15 cents per million input tokens and 60 cents per million output tokens. GPT 4o-mini features In a press release, the Microsoft-backed company said that the new GPT-4o mini can outperform GPT-41 on chat preferences in LMSYS leaderboard.It scored 82% on Massive Multitask Language Understanding (MMLU), OpenAI said. OpenAI's GPT-4o mini enables a broad range of tasks, such as applications that chain or parallelize multiple model calls (e.g., calling multiple APIs), pass a large volume of context to the model (e.g., full code base or conversation history), or interact with customers through fast, real-time text responses (e.g., customer support chatbots). Currently, GPT-4o mini supports text and vision in the API. The company plans to add support for text, image, video and audio inputs and outputs in the future. According to OpenAI, GPT 4o-mini scored 87.0%, compared to 75.5% for Gemini Flash and 71.7% for Claude Haiku in math reasoning. Similarly, GPT-4o mini scored 87.2% on HumanEval, which measures coding performance, compared to 71.5% for Gemini Flash and 75.9% for Claude Haiku. Availability and pricing GPT-4o mini is available as a text and vision model in the Assistants API, Chat Completions API, and Batch API. Developers pay 15 cents per 1M input tokens and 60 cents per 1M output tokens (roughly the equivalent of 2500 pages in a standard book). In ChatGPT, Free, Plus and Team users will be able to access GPT-4o mini starting today, in place of GPT-3.5. Enterprise users will also have access starting next week. Safety measures in GPT-4o mini OpenAI says that the GPT-4o mini has the same safety mitigations built-in as GPT-4o. More than 70 external experts in fields like social psychology and misinformation tested GPT-4o to identify potential risks. Insights from these expert evaluations, the company says, have helped improve the safety of both GPT-4o and GPT-4o mini. The TOI Tech Desk is a dedicated team of journalists committed to delivering the latest and most relevant news from the world of technology to readers of The Times of India. TOI Tech Desk's news coverage spans a wide spectrum across gadget launches, gadget reviews, trends, in-depth analysis, exclusive reports and breaking stories that impact technology and the digital universe. Be it how-tos or the latest happenings in AI, cybersecurity, personal gadgets, platforms like WhatsApp, Instagram, Facebook and more; TOI Tech Desk brings the news with accuracy and authenticity.
[18]
OpenAI launches small AI model GPT-4o Mini. What is it and why is it important?
GPT-4o Mini (O stands for Omni), will replace GPT-3.5 Turbo and will be available to use starting today for free along with ChatGPT Plus and Team members. Meanwhile, it will be offered to enterprise users starting next week. Also Read | OpenAI said that GPT-4o Mini is priced at 15 cents per million input token and 60 cents per million output tokens, making it 60% more cheaper than GPT-3.5 Turbo. The model scored 82% on Massive Multitask Language Understanding (MMLU) and outperformed GPT-4 on chat preferences in LMSYS leaderboard. The company also claimed that GPT-4o Mini also comprehensively defeated other small models in reasoning tasks with Gemini Flash only managing a MMLU score of 77.9% and Claude Haiku a score of 73.8%. Announcing the new model in a blog post, OpenAI wrote, "OpenAI is committed to making intelligence as broadly accessible as possible. Today, we're announcing GPT-4o mini, our most cost-efficient small model. We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable." Small models like GPT-4o require low computational power and hence are a more affordable option for devleopers with limited resources who want to use generative AI in their applications. GPT-4o Mini currently support text and vision in application programming interface (API) and support for text, image, video and audio outputs will be made available in the future, OpenAI said. The latest model has a context window of 128K token, which translates to around 95,000 words, and has a cut off date of October 2023. Meanwhile, OpenAI stated that GPT-4o Mini is even more cost effective handling non-English text now owing to the improved tokenizer. 3.6 Crore Indians visited in a single day choosing us as India's undisputed platform for General Election Results. Explore the latest updates here!
[19]
OpenAI unveils cheaper small AI model GPT-4o mini
(Reuters) - ChatGPT maker OpenAI said on Thursday it was launching GPT-4o mini, a cost-efficient small AI model, aimed at making its technology more affordable and less energy intensive, allowing the startup to target a broader pool of customers. Microsoft-backed OpenAI, which is the market leader in the AI software space, has been working to make it cheaper and faster for developers to build applications based on its model, at a time when deep-pocketed rivals like Meta and Google rush to grab a bigger share in the market.
[20]
What is GPT-4o Mini, OpenAI's newest launch?
Even a baby spider is cuter than its adult counterpart, so when OpenAI debuted a mini version of its most powerful model, you'll have to excuse my instinct to say "Awwww!" That's probably not always the right response, instincts be damned. Because there isn't very much about this new mini version of GPT-4o that we'd describe as "cute." OpenAI's new AI model, GPT-4o mini, is the AI startup's latest effort to make its chatbot the biggest in the game -- even if it's tiny. The company says it is its "most cost-efficient small model." "We expect GPT-4o mini will significantly expand the range of applications built with AI by making intelligence much more affordable," OpenAI said in a press statement. OpenAI went on to specify that the tool currently only supports "text and vision in the API," but said that "support for text, image, video and audio inputs and outputs" will be integrated in the future. It did not specify when that integration will take place. "I think GPT-4o Mini really gets at the OpenAI mission of making AI more broadly accessible to people. If we want AI to benefit every corner of the world, every industry, every application, we have to make AI much more affordable," Olivier Godement, who leads the API platform product, told The Verge. On Thursday, most ChatGPT users -- those who use it on a free, plus, or team plan -- will be able to use GPT-4o Mini instead of GPT-3.5 Turbo. Enterprise users will gain access to the mini model next week.
[21]
OpenAI's GPT-4o mini Offers AI Smarts at a Very Low Price
Tech Deals: Anker Nano Power Bank, Pixel Watch 2, Blink Outdoor Cameras, and More! OpenAI is launching a lightweight version of its flagship GPT-4o model. The new miniaturized model, aptly named "GPT-4o mini," will be available for free and paid ChatGPT users later today. Most AI companies, including Google and Meta, offer some kind of lightweight LLM. These tiny models are intended for simple or repetitive tasks. They don't require a ton of processing power, they're relatively inexpensive, and they're pretty fast. So, they're often the best choice for app or software integration. GPT-4o is the most affordable and lightweight model in OpenAI's arsenal. In fact, it's "60% cheaper" than GPT-3.5 Turbo, and it outscores both Gemini 1.5 Flash and Claude 3 Haiku on the MMLU AI benchmark. Exact benchmark scores should be published on Artificial Analysis within the coming days. For those wondering, the "60% cheaper" claim is totally accurate. Developers utilizing GPT-3.5 Turbo currently pay $0.50 for a million input tokens and $1.50 for a million output tokens. Transiting to GPT-4o mini brings per-million token pricing down to $0.15 and $0.60, respectively. That's important because apps that currently offer AI functionality may improve, and developers that previously avoided AI due to cost concerns may be more inclined to join the bandwagon. The new model might also serve as the replacement for GPT-3.5 for ChatGPT users and developers, but that has not yet been confirmed. Average ChatGPT users may see no reason to test GPT-4o mini, which is understandable, since it's currently aimed at developers. However, free and paid ChatGPT users may still get some use out of GPT-4o mini, especially when generating images, proofreading several pages of text, or asking simple questions. A free user may choose the lightweight model to avoid hitting their daily GPT-4o usage cutoff, for example. GPT-4o mini launches this afternoon (July 18th) and will be immediately available to all ChatGPT users. Developers also gain immediate access, though enterprise customers must wait an extra week. Source: OpenAI via TechCrunch
[22]
OpenAI Slashes the Cost of Using Its AI With a "Mini" Model
With competing models -- including many free ones -- flooding the market, OpenAI is announcing a cheaper way to use its AI. OpenAI today announced a cut-price "mini" model that it says will allow more companies and programs to tap into its artificial intelligence. The new model, called GPT-4o mini, is 60 percent cheaper than OpenAI's most inexpensive existing model while offering higher performance, the company says. OpenAI characterizes the move as part of an effort to make AI "as broadly accessible as possible," but it also reflects growing competition among AI cloud providers as well as rising interest in small and free open source AI models. Meta, the social media giant, is expected to debut the largest version of its very capable free offering, Llama 3, next week. "The whole point of OpenAI is to build and distribute AI safely and make it broadly accessible," Olivier Godement, a product manager at OpenAI responsible for the new model tells WIRED. "Making intelligence available at a lower cost is one of the most efficient ways for us to do that." Godement says the company developed a cheaper offering by improving the model architecture and refining the training data and the training regimen. GPT-4o mini outperforms other "small" models on the market in several common benchmarks, OpenAI says. OpenAI has gained a significant foothold in the cloud AI market thanks to interest in tapping into the remarkable capabilities of its chatbot, ChatGPT, which debuted in late 2022. The company lets outsiders access the large language model that powers ChatGPT, called GPT-4o, for a fee. It also offers a less powerful model called GPT-3.5 Turbo for about a tenth of the cost of GPT-4o. The interest in language models triggered by ChatGPT's wild success has prompted competitors to develop similar offerings. Google, a pioneer in AI, has made a major push to build and commercialize a large language model and chatbot under the brand name Gemini. Startups such as Anthropic, Cohere, and AI21 have raised millions to develop and market their own large language models to business customers and developers. Building the highest-performing large language models requires huge financial resources, but some companies have chosen to open source their creations in order to attract developers to their ecosystems. The most prominent open source AI model is Meta's Llama; it can be downloaded and used for free, but its license imposes certain limits on commercial usage. This April, Meta announced Llama 3, its most powerful free model. The company released a small version of the model with eight billion parameters -- a rough measure of a model's portability and complexity -- as well as a more powerful medium-sized, 70 billion parameter version. The medium-sized model is close to OpenAI's best offering on several benchmark scores. Several sources confirmed to WIRED that Meta plans to release the largest version of Llama 3, with 400 billion parameters, on July 23, although they say the release date could change. It is unclear how capable this version of Llama 3 will be, but some companies have turned their attention toward open source AI models because they are cheaper, can be customized, and offer greater control over a model and the data it is fed. Godement concedes that customers' needs are evolving. "What we see more and more from the market is developers and businesses combining small and large models to build the best product experience at the price and the latency that makes sense for them," he says. Godement says OpenAI's cloud offerings provide customers with models that have gone through more security testing than competitors'. He adds that OpenAI could eventually develop models that customers can run on their own devices. "If we see massive demand, we may open that door," he says.
[23]
OpenAI unveils GPT-4o mini -- a smaller, much cheaper multimodal AI model
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More A little more than two months ago, OpenAI released GPT-4o, its newest and most powerful AI model that was the first to be trained by the company natively to handle multimodal inputs and outputs (text, image, audio, and ultimately video) without linking to other models for help. It was the most powerful, publicly available AI model in the world on third-party benchmarks upon release, but was outclassed shortly after by rival Anthropic's Claude 3.5 Sonnet a few weeks later, and the two have been neck-and-neck ever since. But OpenAI isn't stopping there: today, it is announcing a smaller version of that model, GPT-4o mini, which it says is "the most cost-efficient small model in the market," costing developers just $0.15 USD per 1 million tokens a user inputs, and $0.60 for every million they receive back from the model, for third-party apps and services built atop it using OpenAI's application programming interfaces (APIs). It's also far cheaper than GPT-4o, which costs $5.00 for 1 million input tokens and $15 per 1 million output tokens. Tokens, as you'll recall, are the numerical codes that represent semantic units, words, numbers, and other data inside a given large language model (LLM) or small language model (SML) -- the latter which mini appears to be (OpenAI did not release the number of parameters, or connections between artificial neurons, the model has, making it difficult to say how large or small it is, but the mini name clearly gives an indication.) Olivier Godement, OpenAI's Head of Product, API, told VentureBeat in a teleconference interview yesterday that GPT-4o mini is particularly helpful for enterprises, startups and developers "building any agent" from "a customer support agent" to "a financial agent," as those typically perform "many calls back to the API," resulting in a high volume of tokens inputted and outputted by the underlying source model, which can quickly drive up costs. "The cost per intelligence is so good, I expect it's going to be used for all sorts of customer support, software engineering, creative writing, all kinds of tasks," said Godement. "Every time we adopt a new model, there are new cases that pop up, and I think that will be even more the case for GPT-4o mini." The move to launch GPT-4o mini also comes ahead of Meta's reported release of its massive Llama 3 400-billion parameter model expected next week, and seems quite clearly designed to pre-empt that news and cement in developers' minds that OpenAI remains the leader in enterprise-grade AI. 60% cheaper than GPT-3.5 Turbo for developers To put GPT-4o mini's cost into perspective, it's 60% less than GPT-3.5 Turbo, previously the most affordable model among OpenAI's offerings since the release of GPT-4o. At the same time, the model is targeted to be as fast at working as GPT-3.5 Turbo, transmitting around 67 tokens per second. OpenAI is pitching GPT-4o mini as a direct successor to GPT-3.5 Turbo, but a much more capable one, as it also can handle text and vision inputs, unlike GPT-3.5 Turbo, which could only handle text. At some point in the future, OpenAI says GPT-4o mini will also be able to generate imagery and other multimodal outputs including audio and video, as well as accept them as inputs. But for now, only the text and still image/document inputs will be available today. At present, GPT-4o mini outperforms GPT-3.5 Turbo on a range of third-party benchmarks, other comparably classed models such as Google's Gemini 1.5 Flash and Anthropic's Claude 3 Haiku, and even GPT-4 itself on some tasks. Specifically, OpenAI released benchmarks showing that GPT-4o mini scores 82.0% on the Massive Multitask Language Understanding (MMLU) benchmark, which includes multiple choice questions about subjects from math, science, history, and more, versus 77.9% for Gemini Flash and 73.8% for Claude Haiku. Coming to Apple devices this fall as well In addition, Godement told VentureBeat that GPT-4o mini would be available this fall through Apple Intelligence, the new AI service from Apple Inc., for its mobile devices and Mac desktops, timed to coincide with the release of its new iOS 18 software, as part of the partnership between OpenAI and Apple announced at the latter's WWDC event last month. However, the model will still be running on OpenAI cloud servers -- not on device, which would seem to negate one of the advantages of running a small model in the first place, a local inference that is by nature, faster, more secure, and doesn't require a web connection. Yet Godement pointed out that even when connecting to OpenAI cloud servers, the GPT-4o mini model is faster than others available from the company. Moreover, he told VentureBeat that most third-party developers OpenAI worked with were not yet interested in running the company's models locally, as it would require much more intensive setup and computing hardware on their end. However, the introduction of GPT-4o mini raises the possibility that OpenAI developer customers may now be able to run the model locally more cost effectively and with less hardware, so Godement said it was not out of the question that such a solution could one day be provided. Replacing GPT-3.5 turbo in ChatGPT, but not killing it entirely for developers Beginning later today, GPT-4o mini will replace GPT-3.5 Turbo among the options for paying subscribers of ChatGPT including the Plus and Teams plans -- with support for ChatGPT Enterprise coming next week. The model will appear in the drop-down menu on the upper left corner of the web and Mac desktop apps. However, ChatGPT users won't get a price reduction on their paid subscriptions for selecting GPT-4o mini -- only developers building atop the API will benefit from the savings. Yet ChatGPT users will have access to a newer, faster, and more powerful model for tasks than GPT-3.5 Turbo automatically, which is certainly a benefit. OpenAI isn't yet deprecating or phasing out support for GPT-3.5 Turbo in its APIs, as the company doesn't want to force developers to upgrade or to break the apps that are currently built atop this older model. Instead, the company believes that developers will likely naturally migrate quickly en masse to using the new model since it is a significant cost reduction and boost in intelligence and other capabilities. Some developers have already been alpha testing GPT-4o mini, according to Godement, including enterprise expense management and accounts software startup Ramp and the cloud email AI startup Superhuman, and both are said to have reported excellent results. Godement said GPT-4o mini is powering Ramp's automatic receipt categorization and merchant detection features, and powering Superhuman's suggested, custom-tailored email responses. Ramp in particular has "seen pretty amazing results for its data extraction tests," from receipts, said Godement. He was not able to say precisely whether Ramp was using the GPT-4o mini native multimodal vision input or if the firm was using another system to first extract text and numerals from receipts and send it to the model. So why should any developers still use the older, more expensive GPT-4o parent model? Given the significant cost savings offered by GPT-4o mini and high performance benchmarks on a number of tasks and tests, the question naturally arises: why would a developer pay more money to use the full GPT-4o model when the mini one is now available? OpenAI believes that for the most computationally-intensive, complex, and demanding applications, the full GPT-4o is still the way to go, and justifies its higher price in comparison. "Let's assume I'm building medical applications that I'd like to summarize and propose some diagnosis for patients," Godement gave as one example. "I'm basically going to optimize for intelligence. I want to make sure they get the most intelligent model out of the box. Similarly, if you're building a software engineering assistant and working on a pretty complex codebase, you will still be see better results with GPT-4o. If intelligence differentiates your product, I recommend you stick with GPT-4o and you'll get the best results."
[24]
OpenAI, Mistral AI debut new cost-efficient language models - SiliconANGLE
OpenAI and Mistral AI today introduced new language models for powering applications that must balance output quality with cost-efficiency. OpenAI's new model, GPT-4o mini, is a scaled-down version of its flagship GPT-4o large language model. Mistral AI, in turn, debuted an algorithm dubbed Mistral NeMo 12B that was developed in collaboration with Nvidia Corp. engineers. It's designed for many of the same tasks as GPT-4o mini and will be available under an open-source license. GPT-4o mini can generate text, craft code and solve math problems much like its more capable namesake. However, the model does so slightly less accurately. GPT-4o mini achieved a score of 82% on MMLU, a benchmark test used to measure the quality of language models' output, while the original GPT-4o scored 88.7%. GPT-4o mini trades off those few percentage points' worth of accuracy for increased cost-efficiency. It will be available through OpenAI's application programming interface for less than a fifth of the price at which the company offers GPT-4o. As a result, the applications that developers build using the API will be less expensive to operate. The model is the first from OpenAI to include a technology called instruction hierarchy. The feature, which the company first detailed in an April research paper, is designed to reduce the risk posed by malicious user input. Services powered by an OpenAI-developed model such as GPT-4o mini often receive multiple types of prompts. There are prompts entered by an application's developer that might, for example, instruct GPT-4o mini not to disclose sensitive data to users. Separately, the application's users send their own requests to the model. Instruction hierarchy blocks malicious input by prioritizing the developer's prompts over the ones entered by users. If a developer instructs an application powered by GPT-4o mini not to disclose sensitive data but a user asks the model to do so regardless, the model will reject the latter request. It can also prioritize developer-provided instructions in other situations to present applications from carrying out tasks they were not intended to perform. GPT-4o mini will become available via OpenAI's API next week. The company is also rolling out GPT-4o mini to all four tiers of its ChatGPT chatbot service, including the free plan and the top-end Enterprise edition. The latter offering will receive a number of other new features as well through a separate update detailed this morning. Companies in regulated industries such as the healthcare sector must often keep a record of internal business activities. As part of today's update, ChatGPT Enterprise is receiving an API that will enable organizations to download a log of their employees' interactions with the service. OpenAI is also rolling out features that will make it easier to create and delete employee accounts, as well as a tool for blocking integrations with unauthorized third-party applications. Like GPT-4o mini, the open-source Mistral NeMo 12B model that Mistral AI and Nvidia debuted today is designed to be more cost-efficient than frontier LLMs. It features 12 billion parameters, the configuration settings that determine how a neural network processes data. That's significantly less than the hundreds of millions of parameters in frontier LLMs, which means Mistral NeMo 12B can perform inference using less hardware and thereby cut users' infrastructure costs. Nvidia detailed in a blog post that the model is compact enough to run in the memory of a single graphics processing unit. The GeForce RTX 4090, a high-end consumer GPU that the chipmaker debuted in 2022, is among the chips that can accommodate Mistral NeMo 12B. Customers can also run the model on a single RTX 4500, which is designed to power workstations, or Nvidia's entry-level L40S data center graphics card. The chipmaker and Mistral AI have packaged Mistral NeMo 12B into a so-called NIM microservice. That's a preconfigured software container designed to ease the task of deploying the model on Nvidia silicon. According to the chipmaker, NIM can reduce the amount of time required to deploy a neural network from days to a few minutes. Mistral AI and Nvidia envision developers using Mistral NeMo 12B to power chatbot services. The model also lends itself to several other tasks including code generation, translation and documentation summarization.
Share
Share
Copy Link
OpenAI introduces GPT-4o Mini, a smaller and more affordable version of GPT-4. This new AI model aims to reduce costs for developers while maintaining impressive capabilities.
In a significant move that could reshape the AI landscape, OpenAI has unveiled its latest artificial intelligence model, GPT-4o Mini. This new offering is designed to be a more compact and cost-effective version of its predecessor, GPT-4, while still delivering powerful performance 1.
GPT-4o Mini boasts impressive capabilities despite its smaller size. The model can process up to 8,000 tokens, which is equivalent to about 6,000 words or 16 pages of text 2. This capacity allows it to handle a wide range of tasks, from text generation to complex problem-solving.
One of the standout features of GPT-4o Mini is its ability to generate human-like text based on prompts. This makes it particularly useful for applications such as chatbots, content creation, and language translation 3.
The primary goal behind GPT-4o Mini is to make advanced AI technology more accessible to developers and businesses. OpenAI has priced the model at a fraction of the cost of GPT-4, with rates starting at $0.01 per 1,000 tokens for input and $0.03 per 1,000 tokens for output 4.
This pricing strategy is expected to significantly reduce the cost of running AI applications, potentially leading to wider adoption across various industries 5.
The introduction of GPT-4o Mini is likely to have far-reaching implications for the AI industry. By offering a more affordable option, OpenAI is potentially democratizing access to advanced AI capabilities. This could lead to an increase in AI-powered applications and services across various sectors 3.
Moreover, the launch of GPT-4o Mini could intensify competition in the AI market. Other major players in the field may feel pressure to develop similar cost-effective models to maintain their market share 5.
While GPT-4o Mini represents a significant advancement in making AI more accessible, it also raises questions about the future direction of AI development. The balance between model size, performance, and cost will likely continue to be a key area of focus for researchers and developers in the field 1.
As AI models become more compact and affordable, there may also be increased scrutiny regarding their ethical use and potential misuse. Ensuring responsible AI deployment will remain a crucial consideration as these technologies become more widely available 4.
Reference
[2]
[4]
[5]
OpenAI has introduced GPT-4O Mini, a new AI model designed to be faster and more cost-effective than its predecessors. This lightweight version aims to replace GPT-3.5 and offers improved performance for ChatGPT users.
10 Sources
OpenAI has introduced GPT-4o Mini, a more affordable version of its top AI model. This new offering aims to make advanced AI technology more accessible to developers and businesses while potentially reshaping the competitive landscape in the AI industry.
5 Sources
The AI industry is witnessing a shift in focus from larger language models to smaller, more efficient ones. This trend is driven by the need for cost-effective and practical AI solutions, challenging the notion that bigger models are always better.
2 Sources
OpenAI has introduced its latest AI model series, O1, featuring enhanced reasoning abilities and specialized variants. While showing promise in various applications, the models also present challenges and limitations.
5 Sources
OpenAI's DevDay 2024 unveiled groundbreaking updates to its API services, including real-time voice interactions, vision fine-tuning, prompt caching, and model distillation techniques. These advancements aim to enhance developer capabilities and unlock new possibilities in AI-powered applications.
5 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2024 TheOutpost.AI All rights reserved