Curated by THEOUTPOST
On Mon, 10 Mar, 4:03 PM UTC
9 Sources
[1]
Foxconn launches traditional Chinese large language model for AI-driven manufacturing
Foxconn's logo is displayed during the Hon Hai Tech Day at the Nangang Exhibition Center in Taipei, Taiwan, Oct. 8, 2024. AP-Yonhap Foxconn Technology Group, the world's largest electronics contract manufacturer and major iPhone supplier for Apple, launched its first Chinese large language model (LLM) trained on traditional characters, as the Taiwanese company pushes forward the use of artificial intelligence (AI) in factories. The new FoxBrain model was trained in a "more efficient and lower-cost" method within just four weeks, and sets a new milestone in the development of Taiwan's AI technology, according to a statement issued on Monday by Foxconn, known formally as Hon Hai Precision Industry. With a training process powered by 120 Nvidia H100 graphics processing units (GPUs), FoxBrain excels in math and logical reasoning, according to Foxconn. It was originally designed for internal applications in the company, but Foxconn said it will be open sourced in the future, as part of efforts to collaborate with technology partners to expand its applications and promote AI in manufacturing. The Nvidia's GPU (Graphic Processing Unit) is shown in this photo taken in Paris, Febr. 23, 2024. AFP-Yonhap LLMs are the technology underpinning generative AI services like OpenAI's ChatGPT. Open source gives public access to a software's source code, allowing third-party developers to modify or share its design, fix broken links or scale up its capabilities. Foxconn's latest initiative reflects the company's goal to push its own AI breakthroughs in terms of manufacturing efficiency. That follows Chinese start-up DeepSeek's launch earlier this year of its high-performance R1 reasoning model, which was open-sourced and developed at a fraction of the cost of AI models from larger companies like OpenAI, Google and Meta Platforms. "In recent months, the deepening of reasoning capabilities and the efficient use of GPUs have gradually become the mainstream development in the field of AI," said Li Yung-Hui, director of the Artificial Intelligence Research Centre at Hon Hai Research Institute, the research arm of the manufacturer, in the statement. "Our FoxBrain model adopted a very efficient training strategy, focusing on optimizing the training process rather than blindly accumulating computing power," he said. The new model was based on the Meta Llama 3.1 architecture with 70 billion parameters. Foxconn claimed that it outperformed Llama-3-Taiwan-70B, another open-source model fine-tuned on traditional Chinese characters and English data using the Llama-3 architecture, in most categories of TMMLU+, a benchmark for traditional Chinese language understanding. Models developed by Chinese companies like DeepSeek are generally trained for better understanding of simplified Chinese characters, which are used on the mainland. Last November, the company said it was working with Nvidia to leverage "digital twin" technology in manufacturing and supply chain management. The initiative uses Nvidia's Omni verse to streamline global factory operations, enhance resilience and ensure consistent quality. Read the full story at SCMP.
[2]
FoxBrain AI Model Unveiled by Foxconn, Built in Just Four Weeks
Foxconn wants to open-source the FoxBrain model to drive advancements in manufacturing and supply-chain management. Foxconn, a company based out of Taiwan, is one of the world's largest electronics manufacturers. It's one of the key companies that assemble the iPhone, MacBook, PlayStation, and more. Now, Foxconn has announced that it has trained its own AI model called 'FoxBrain'. It's a large language model just like OpenAI's GPT series, Google's Gemini, and Anthropic's Claude. Foxconn says the FoxBrain AI model can perform a variety of tasks including data analysis, code generation, reasoning, and mathematical calculations. It's based on Meta's Llama 3.1 model, however, FoxBrain also exhibits advanced reasoning capabilities. On top of that, the Foxconn-trained model is optimized for the Chinese language. The surprising part is that Foxconn managed to train the AI model in just four weeks, thanks to support from Nvidia. Nvidia offered its Taiwan-based supercomputer to Foxconn for training the model. In addition, the chipmaker also offered technical consulting for model training. As for hardware requirements for training, Foxconn utilized 120 Nvidia H100 GPUs. Foxconn says the FoxBrain AI model can be used to improve manufacturing and supply chain management. Moreover, the company is keen to open-source the model so that industry partners can collaborate and make advancements in the manufacturing sector. Note that Foxconn has not revealed the parameter count or benchmarks for the FoxBrain AI model.
[3]
Apple Supplier Foxconn Unveils AI Model Using Nvidia Tech - Foxconn Tech Co (OTC:FXCOF)
On Monday, key Apple Inc AAPL supplier Foxconn told the Wall Street Journal it has built its in-house large language model with reasoning capabilities and trained in four weeks. Foxconn's profitability challenges in its core electronics manufacturing business prompted it to diversify into artificial intelligence and electric vehicles. The AI model, called FoxBrain, is capable of data analysis, mathematics, reasoning, and code generation. Also Read: ChatGPT Parent OpenAI's Specialized AI Agents Could Cost Up To $20 Thousand Per Month: report Foxconn expects FoxBrain to drive advancements in manufacturing and supply chain management. Foxconn said 120 Nvidia Corp NVDA H100 graphics processing units provided support through its Taiwan-based supercomputer and technical consulting. Foxconn plans to open-source the model for collaborations with industry partners. FoxBrain is based on the structure of Meta Platforms Inc's META publicly available large language model Llama 3.1. Global tech companies remain focused on introducing AI models to gain market share amid the AI frenzy. Amazon.Com Inc AMZN plans to launch AI reasoning model in June 2025 for advanced problem-solving capabilities. Meta plans to introduce improved voice features into its latest open-source large language model, Llama 4. The Trump administration is reportedly considering restricting Chinese AI startup DeepSeek, including banning its chatbot from government devices shortly after DeepSeek made its AI models open-source. Earlier in 2025, DeepSeek's claims of its AI model built at a fraction of the cost of its U.S. rivals cost Nvidia $600 billion in market cap in a single day. DeepSeek's AI model followed a $1 trillion market wipeout. Also Read: USA Targets DeepSeek, Weighs Restrictions On China's AI Firm That Shook Nvidia's Market Value By $600 Billion Image by Pixels Hunter via Shutterstock FXCOFFoxconn Technology Co Ltd$3.0286.3%Stock Score Locked: Want to See it? Benzinga Rankings give you vital metrics on any stock - anytime. Reveal Full ScoreEdge RankingsMomentum-Growth-Quality-Value-Price TrendShortMediumLongOverviewAAPLApple Inc$225.28-5.77%AMZNAmazon.com Inc$193.29-2.99%METAMeta Platforms Inc$594.77-4.94%NVDANVIDIA Corp$107.07-4.99%Market News and Data brought to you by Benzinga APIs
[4]
Foxconn Unveils FoxBrain -- Chinese AI Model Poised for Open Source Release
FoxBrain's training process with NVIDIA GPUs finished in just four weeks. Hon Hai Research Institute, a research centre by Hon Hai Technology Group (Foxconn), launched a traditional Chinese large language model (LLM) code-named FoxBrain on Monday. FoxBrain was initially designed for internal group applications, but it is set to become an important engine for Foxconn's three major platforms: Smart Manufacturing, Smart EV, and Smart City. It was initially developed for the Group's internal systems, supporting applications in data analysis, decision support, document collaboration, mathematics, reasoning, problem-solving, and code generation. The institute also mentioned that the model will be open-sourced and shared publicly in the future. The institute noted that the development was a milestone in Taiwan's AI technology, where an efficient and lower-cost model training method was completed in just four weeks. FoxBrain's quick training process was made possible using 120 NVIDIA H100 GPUs, scaled with NVIDIA Quantum-2 InfiniBand network. NVIDIA helped during the training process through the Taipei-1 Supercomputer and technical consultation. The model is based on Meta's Llama 3.1 architecture with 70B parameters and features 128k tokens for its context window length. It is said that a unique adaptive reasoning reflection technique was used to train the model in autonomous reasoning. "In recent months, the deepening of reasoning capabilities and the efficient use of GPUs have gradually become the mainstream development in the field of AI. Our FoxBrain model adopted a very efficient training strategy, focusing on optimising the training process rather than blindly accumulating computing power," said Yung-Hui Li, director of the Artificial Intelligence Research Centre at the Hon Hai Research Institute. "Through carefully designed training methods and resource optimisation, we have successfully built a local AI model with powerful reasoning capabilities." The institute mentioned that its performance is close to world-leading standards, but there is a slight gap with DeepSeek's distillation model. The institute stated that it outperforms Llama-3-Taiwan-70B on the same scale, with mathematics and logical reasoning as the focus. However, the result of FoxBrain is scheduled to be shared at NVIDIA GTC 2025 in a session talk titled 'From Open Source to Frontier AI: Build, Customise, and Extend Foundational Models' on March 20.
[5]
Foxconn unveils AI model FoxBrain for improving manufacturing processes By Investing.com
The company plans to use this technology to enhance manufacturing and supply chain management processes. FoxBrain was developed using 120 of Nvidia (NASDAQ:NVDA)'s H100 GPUs and was completed in approximately four weeks. The model is built on the Llama 3.1 architecture, a technology developed by Meta (NASDAQ:META). Although Foxconn acknowledged a slight performance gap when compared with China's DeepSeek's distillation model, it stated that FoxBrain's overall performance is very close to world-class standards. In order to expand the applications of FoxBrain, Foxconn plans to collaborate with technology partners, share its open-source information, and promote the use of artificial intelligence in areas such as manufacturing, supply chain management, and intelligent decision-making.
[6]
Foxconn unveils first large language model
TAIPEI, March 10 (Reuters) - Taiwan's Foxconn (2317.TW), opens new tab said on Monday it has launched its first large language model and plans to use the technology to improve manufacturing and supply chain management. The model, named "FoxBrain," was trained using 120 of Nvidia's (NVDA.O), opens new tab H100 GPUs and completed in about four weeks, the world's largest contract electronics manufacturer said in a statement. The company, which assembles iPhones for Apple (AAPL.O), opens new tab and also produces Nvidia's artificial intelligence servers, said the model is based on Meta's (META.O), opens new tab Llama 3.1 architecture. It is Taiwan's first large language model with reasoning capabilities that is optimised for traditional Chinese and Taiwanese language styles, it said. Foxconn said that though there was a slight performance gap compared with China's DeepSeek's distillation model, its overall performance is very close to world-class standards. Initially designed for internal applications, FoxBrain covers data analysis, decision support, document collaboration, mathematics, reasoning and problem-solving, and code generation. Foxconn said it plans to collaborate with technology partners to expand the model's applications, share its open-source information, and promote AI in manufacturing, supply chain management, and intelligent decision-making. Nvidia provided support through its Taiwan-based supercomputer "Taipei-1" and offered technical consulting during the model's training, Foxconn said. Taipei-1, the largest supercomputer in Taiwan, is owned and operated by Nvidia in Kaohsiung, a southern city on the island. Foxconn will announce further details about the model during Nvidia's GTC developer conference in mid-March. Reporting by Wen-Yee Lee Editing by Shri Navaratnam Our Standards: The Thomson Reuters Trust Principles., opens new tab Suggested Topics:Disrupted
[7]
Foxconn Unveils First Large Language Model Dubbed 'FoxBrain'
FoxBrain covers data analysis, decision support, document collaboration Taiwan's Foxconn said on Monday it has launched its first large language model and plans to use the technology to improve manufacturing and supply chain management. The model, named "FoxBrain," was trained using 120 of Nvidia's H100 GPUs and completed in about four weeks, the world's largest contract electronics manufacturer said in a statement. The company, which assembles iPhones for Apple and also produces Nvidia's artificial intelligence servers, said the model is based on Meta's Llama 3.1 architecture. It is Taiwan's first large language model with reasoning capabilities that is optimised for traditional Chinese and Taiwanese language styles, it said. Foxconn said that though there was a slight performance gap compared with China's DeepSeek's distillation model, its overall performance is very close to world-class standards. Initially designed for internal applications, FoxBrain covers data analysis, decision support, document collaboration, mathematics, reasoning and problem-solving, and code generation. Foxconn said it plans to collaborate with technology partners to expand the model's applications, share its open-source information, and promote AI in manufacturing, supply chain management, and intelligent decision-making. Nvidia provided support through its Taiwan-based supercomputer "Taipei-1" and offered technical consulting during the model's training, Foxconn said. Taipei-1, the largest supercomputer in Taiwan, is owned and operated by Nvidia in Kaohsiung, a southern city on the island. Foxconn will announce further details about the model during Nvidia's GTC developer conference in mid-March. © Thomson Reuters 2025
[8]
Foxconn Unveils Its First LLM FoxBrain to Optimise Manufacturing
Future plans include open-source availability and broader AI applications in smart manufacturing, EVs, and smart cities. Taiwan's Foxconn has launched its first Large Language Model (LLM), FoxBrain, to enhance manufacturing and supply chain operations. Based on Meta's Llama 3.1, the AI model features 70 billion parameters and a 128k-token context window. It was trained in four weeks using 120 Nvidia H100 GPUs, scaled with Nvidia Quantum-2 InfiniBand networking, and employs adaptive reasoning reflection to improve efficiency. Also Read: Tech Mahindra Launches AI Network Automation Model for Telcos With Nvidia and AWS "The institute [Hon Hai Research Institute], backed by Hon Hai Technology Group (Foxconn), the world's largest electronics manufacturer and technological solutions provider, said the LLM - code-named FoxBrain - will be open-sourced and shared publicly in the future. It was originally designed for applications used in the Group's internal systems, covering functions such as data analysis, decision support, document collaboration, mathematics, reasoning and problem-solving, and code generation." Foxconn said on March 10, 2025. "In recent months, the deepening of reasoning capabilities and the efficient use of GPUs have gradually become the mainstream development in the field of AI. Our FoxBrain model adopted a very efficient training strategy, focusing on optimising the training process rather than blindly accumulating computing power," said Yung-Hui Li, Director of the Artificial Intelligence Research Center at Hon Hai Research Institute. "Through carefully designed training methods and resource optimisation, we have successfully built a local AI model with powerful reasoning capabilities." Also Read: IBM and L'Oreal to Build AI Model for Creation of Sustainable Cosmetics Foxconn claims that FoxBrain outperforms Llama-3-Taiwan-70B in math and logical reasoning but still lags behind DeepSeek's model. "In test results, FoxBrain showed comprehensive improvements in mathematics compared to the base Meta Llama 3.1 model. It achieved significant progress in mathematical tests compared to Taiwan Llama, currently the best Traditional Chinese large model, and surpassed Meta's current models of the same class in mathematical reasoning ability. While there is still a slight gap with DeepSeek's distillation model, its performance is already very close to world-leading standards," Foxconn added. The company stated that this large language model research demonstrates Taiwan's technological talent can compete with international counterparts in the AI field. Foxconn also announced that, in the future, the Group will continue collaborating with technology partners to expand FoxBrain's applications, share its open-source information, and promote AI in manufacturing, supply chain management, and intelligent decision-making. Also Read: Meta Unveils New AI Models and Tools to Drive Innovation During model training, Foxconn said Nvidia provided support through the Taipei-1 Supercomputer and technical consultation, enabling Hon Hai Research Institute to complete the model pre-training with Nvidia NeMo. "FoxBrain will also become an important engine to drive the upgrade of Foxconn's three major platforms: Smart Manufacturing. Smart EV. Smart City," the official release said. The company will present its results at Nvidia GTC 2025 on March 20.
[9]
Foxconn unveils first large language model
TAIPEI (Reuters) - Taiwan's Foxconn said on Monday it has launched its first large language model and plans to use the technology to improve manufacturing and supply chain management. The model, named "FoxBrain," was trained using 120 of Nvidia's H100 GPUs and completed in about four weeks, the world's largest contract electronics manufacturer said in a statement. The company, which assembles iPhones for Apple and also produces Nvidia's artificial intelligence servers, said the model is based on Meta's Llama 3.1 architecture. It is Taiwan's first large language model with reasoning capabilities that is optimised for traditional Chinese and Taiwanese language styles, it said. Foxconn said that though there was a slight performance gap compared with China's DeepSeek's distillation model, its overall performance is very close to world-class standards. Initially designed for internal applications, FoxBrain covers data analysis, decision support, document collaboration, mathematics, reasoning and problem-solving, and code generation. Foxconn said it plans to collaborate with technology partners to expand the model's applications, share its open-source information, and promote AI in manufacturing, supply chain management, and intelligent decision-making. Nvidia provided support through its Taiwan-based supercomputer "Taipei-1" and offered technical consulting during the model's training, Foxconn said. Taipei-1, the largest supercomputer in Taiwan, is owned and operated by Nvidia in Kaohsiung, a southern city on the island. Foxconn will announce further details about the model during Nvidia's GTC developer conference in mid-March. (Reporting by Wen-Yee Lee; Editing by Shri Navaratnam)
Share
Share
Copy Link
Foxconn, a major electronics manufacturer, has introduced FoxBrain, an AI large language model developed in just four weeks. The model aims to enhance manufacturing processes and will be open-sourced for industry collaboration.
Foxconn Technology Group, the world's largest electronics contract manufacturer and a major supplier for Apple, has unveiled its first Chinese large language model (LLM) called FoxBrain. This AI model, trained on traditional Chinese characters, marks a significant step in the company's push towards AI-driven manufacturing 1.
FoxBrain was developed in a remarkably short timeframe of just four weeks, using what Foxconn describes as a "more efficient and lower-cost" method. The training process utilized 120 Nvidia H100 graphics processing units (GPUs), highlighting the computational power behind this AI model 2.
Based on Meta's Llama 3.1 architecture with 70 billion parameters, FoxBrain boasts a context window length of 128k tokens. The model excels in areas such as math, logical reasoning, data analysis, and code generation. Foxconn claims that FoxBrain outperforms Llama-3-Taiwan-70B in most categories of TMMLU+, a benchmark for traditional Chinese language understanding 4.
FoxBrain was initially designed for internal applications within Foxconn but is set to become a crucial engine for the company's three major platforms: Smart Manufacturing, Smart EV, and Smart City. The company expects FoxBrain to drive advancements in manufacturing processes, supply chain management, and intelligent decision-making 3.
In a move to foster innovation and collaboration, Foxconn has announced plans to open-source the FoxBrain model in the future. This decision aims to expand its applications and promote AI adoption in manufacturing by allowing industry partners to modify, share, and scale up its capabilities 5.
While Foxconn acknowledges a slight performance gap compared to China's DeepSeek's distillation model, it maintains that FoxBrain's overall performance is close to world-leading standards. The company plans to share more detailed results at the NVIDIA GTC 2025 conference 4.
The development of FoxBrain reflects Foxconn's broader strategy to diversify beyond its core electronics manufacturing business. Faced with profitability challenges, the company is expanding into artificial intelligence and electric vehicles to maintain its competitive edge in the rapidly evolving tech landscape 3.
Reference
[1]
[4]
Analytics India Magazine
|Foxconn Unveils FoxBrain -- Chinese AI Model Poised for Open Source ReleaseFoxconn is constructing a massive production facility in Mexico for Nvidia's next-generation Blackwell AI chips, signaling a significant shift in the global AI hardware manufacturing landscape and Foxconn's diversification beyond consumer electronics.
11 Sources
11 Sources
Foxconn and NVIDIA are collaborating to construct Taiwan's most powerful AI supercomputer, utilizing NVIDIA's Blackwell architecture. The project aims to advance AI research, healthcare, and smart city innovations.
6 Sources
6 Sources
Foxconn, a major Apple supplier, reports a 6% increase in second-quarter net profit, driven by growing demand for artificial intelligence servers. The company's performance exceeds market expectations despite challenges in the consumer electronics sector.
11 Sources
11 Sources
Foxconn reports record Q4 2024 revenue, citing strong AI server demand. This news has led to a surge in semiconductor stocks, highlighting the growing impact of AI on the tech industry.
11 Sources
11 Sources
Apple is reportedly in discussions with Foxconn and Lenovo to manufacture AI servers using Apple Silicon, aiming to power its Apple Intelligence services and boost its AI capabilities.
4 Sources
4 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved