Curated by THEOUTPOST
On Wed, 18 Dec, 12:01 AM UTC
2 Sources
[1]
UAE's TII Announces 'Powerful' Small AI Model Falcon 3
The small model is available in multiple variants up to 10 billion parameters, and has been released on Hugging Face. Technology Innovation Institute (TII), a research institute from Abu Dhabi, UAE, has unveiled a new family of small language models titled Falcon 3. The models range from one billion parameters to 10 billion parameters in both base and instruct versions. Falcon is available as an open-source model under TII's Falcon License 2.0. The institute also released benchmark results comparing some of the other leading models in its category. Both Falcon 3 7B and 10B outperformed models like Qwen 2.5 7B and Llama 3.1 8B in several benchmarks. TII is a global research institution based in Abu Dhabi and funded by the Abu Dhabi government. It was established in May 2020 and focuses on research in AI, quantum computing, robotics, and cryptography. Falcon 3 employed a shared parameter technique called Grouped Query Attention (GQA) that reduces the memory demands, thereby leading to low latency during inference. "The initial training was followed by multiple stages to improve reasoning and math performance with high-quality data and context extension with natively long context data," read the announcement. The model was also trained in four languages, including English, Spanish, Portuguese, and French. All variants of the Falcon model are available for download on Hugging Face. In August, TII launched the Falcon Mamba 7B model. It outperformed Meta's Llama 3.1 8B, Llama 3 8B, and Mistral's 7B in benchmarks. In May, they launched Falcon 2, an 11B text and vision model. Are small language models finally delivering the promise? A few days ago, Microsoft announced the latest Phi-4 model. With just 14B parameters, the model outperformed much larger models like Llama 3.3 70B and GPT 4o on several benchmarks. There have also been discussions about the relevance of pre-training and a brute-force approach to improve the model by increasing its size. Ilya Sutskever, former OpenAI chief scientist, had his say on this debate in his presentation at NeurIPS 2024. "Pre-training as we know it will unquestionably end," he said, referring to the lack of available data. "We have but one internet. You could even go as far as to say that data is the fossil fuel of AI. It was created somehow, and now we use it," he added. He also speculated that the use of inference time computing and synthetic data for training is a key technique that may help researchers overcome the problem. That said, if small models leverage new and innovative techniques and deliver high performance on resource-constrained devices, the smartphone market in 2025 will be the one to watch out for.
[2]
UAE's Falcon 3 challenges open-source leaders amid surging demand for small AI models
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More The UAE government-backed Technology Innovation Institute (TII) has announced the launch of Falcon 3, a family of open-source small language models (SLMs) designed to run efficiently on lightweight, single GPU-based infrastructures. Falcon 3 features four model sizes -- 1B, 3B, 7B, and 10B -- with base and instruct variants, promising to democratize access to advanced AI capabilities for developers, researchers, and businesses. According to the Hugging Face leaderboard, the models are already outperforming or closely matching popular open-source counterparts in their size class, including Meta's Llama and category leader Qwen-2.5. The development comes at a time when the demand for SLMs, with fewer parameters and simpler designs than LLMs, is rapidly growing due to their efficiency, affordability, and ability to be deployed on devices with limited resources. They are suitable for a range of applications across industries, like customer service, healthcare, mobile apps and IoT, where typical LLMs might be too computationally expensive to run effectively. According to Valuates Reports, the market for these models is expected to grow, with a CAGR of nearly 18% over the next five years. What does Falcon 3 bring to the table? Trained on 14 trillion tokens -- more than double its predecessor Falcon 2 -- the Falcon 3 family employs a decoder-only architecture with grouped query attention to share parameters and minimize memory usage for key-value (KV) cache during inference. This enables faster and more efficient operations when handling diverse text-based tasks. At the core, the models support four primary languages -- English, French, Spanish, and Portuguese -- and come equipped with a 32K context window, allowing them to process long inputs, such as heavily worded documents. "Falcon 3 is versatile, designed for both general-purpose and specialized tasks, providing immense flexibility to users. Its base model is perfect for generative applications, while the instruct variant excels in conversational tasks like customer service or virtual assistants," TII notes on its website. According to the leaderboard on Hugging Face, while all four Falcon 3 models perform fairly well, the 10B and 7B versions are the stars of the show, achieving state-of-the-art results on reasoning, language understanding, instruction following, code and mathematics tasks. Among models under the 13B-parameter size class, Falcon 3's 10B and 7B versions outperform competitors, including Google's Gemma 2-9B, Meta's Llama 3.1-8B, Mistral-7B, and Yi 1.5-9B. They even surpass Alibaba's category leader Qwen 2.5-7B in most benchmarks -- such as MUSR, MATH, GPQA, and IFEval -- except for MMLU, which is the test for evaluating how well language models understand and process human language. Deployment across industries With the Falcon 3 models now available on Hugging Face, TII aims to serve a broad range of users, enabling cost-effective AI deployments without computational bottlenecks. With their ability to handle specific, domain-focused tasks with fast processing times, the models can power various applications at the edge and in privacy-sensitive environments, including customer service chatbots, personalized recommender systems, data analysis, fraud detection, healthcare diagnostics, supply chain optimization and education. The institute also plans to expand the Falcon family further by introducing models with multimodal capabilities. These models are expected to launch sometime in January 2025. Notably, all models have been released under the TII Falcon License 2.0, a permissive Apache 2.0-based license with an acceptable use policy that encourages responsible AI development and deployment. To help users get started, TII has also launched a Falcon Playground, a testing environment where researchers and developers can try out Falcon 3 models before integrating them into their applications.
Share
Share
Copy Link
The Technology Innovation Institute (TII) in UAE has launched Falcon 3, a family of small language models ranging from 1B to 10B parameters, outperforming larger models in various benchmarks and promising efficient AI deployment across industries.
The Technology Innovation Institute (TII), a research institute based in Abu Dhabi, UAE, has unveiled Falcon 3, a groundbreaking family of small language models (SLMs) that are set to challenge the dominance of larger AI models 1. This development marks a significant step forward in the field of artificial intelligence, particularly in the realm of efficient and accessible AI technologies.
Falcon 3 comes in multiple variants, ranging from 1 billion to 10 billion parameters, available in both base and instruct versions 1. The models have been trained on an impressive 14 trillion tokens, more than doubling the training data of its predecessor, Falcon 2 2. Key features of Falcon 3 include:
In benchmark tests, Falcon 3's 7B and 10B variants have outperformed several leading models in their category, including Qwen 2.5 7B, Llama 3.1 8B, and even surpassed Alibaba's Qwen 2.5-7B in most benchmarks except MMLU 12.
TII has made all variants of Falcon 3 available for download on Hugging Face under the TII Falcon License 2.0, an Apache 2.0-based permissive license 12. This move aligns with the growing trend of democratizing access to advanced AI capabilities, allowing developers, researchers, and businesses to leverage these powerful models.
The introduction of Falcon 3 comes at a time when demand for SLMs is rapidly growing due to their efficiency and affordability 2. These models are particularly suited for applications that require:
Potential applications span various industries, including:
Falcon 3's release contributes to the ongoing debate about the future of AI model development. Recent developments, such as Microsoft's Phi-4 model, have demonstrated that smaller models can outperform much larger ones on several benchmarks 1. This trend challenges the notion that increasing model size is the primary path to improved performance.
Ilya Sutskever, former OpenAI chief scientist, recently commented on this shift at NeurIPS 2024, stating, "Pre-training as we know it will unquestionably end," and suggesting that techniques like inference time computing and synthetic data for training may be key to overcoming data limitations 1.
TII has announced plans to expand the Falcon family further by introducing models with multimodal capabilities, expected to launch in January 2025 2. To facilitate adoption and experimentation, TII has also launched a Falcon Playground, allowing researchers and developers to test the models before integration into their applications 2.
As the AI landscape continues to evolve, the success of Falcon 3 and similar SLMs could significantly impact the smartphone market in 2025, potentially revolutionizing on-device AI capabilities 1. This development represents a crucial step towards more accessible, efficient, and powerful AI technologies across various sectors.
Reference
[1]
Mistral AI unveils Mistral Small 3, a 24-billion-parameter open-source AI model that rivals larger competitors in performance while offering improved efficiency and accessibility.
4 Sources
4 Sources
Microsoft has released a new series of Phi-3.5 AI models, showcasing impressive performance despite their smaller size. These models are set to compete with offerings from OpenAI and Google, potentially reshaping the AI landscape.
4 Sources
4 Sources
Hugging Face introduces SmolVLM-256M and SmolVLM-500M, the world's smallest vision-language AI models capable of running on consumer devices while outperforming larger counterparts, potentially transforming AI accessibility and efficiency.
5 Sources
5 Sources
Google has released updated versions of its Gemma large language models, focusing on improved performance, reduced size, and enhanced safety features. These open-source AI models aim to democratize AI development while prioritizing responsible use.
2 Sources
2 Sources
Recent developments suggest open-source AI models are rapidly catching up to closed models, while traditional scaling approaches for large language models may be reaching their limits. This shift is prompting AI companies to explore new strategies for advancing artificial intelligence.
5 Sources
5 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved