Taiwan’s Economic Daily reported on March 10 that Foxconn’s Hon Hai Research Institute (HHRI) announced the development of FoxBrain, Taiwan's first internally developed traditional Chinese large language model (LLM). This model is optimized for Taiwanese language styles and user interaction patterns, showing computational efficiency with advanced reasoning capabilities comparable to DeepSeek’s V3, the Chinese LLM of R1, which has made a wave lately.
FoxBrain is a 70B-parameter LLM based on Meta Llama 3.1, optimized for traditional Chinese and tailored to Taiwanese linguistic and cultural nuances. Initially designed for internal applications, it supports tasks like data analysis, decision-making assistance, code generation, and clerical automation. HHRI plans to share the model externally to expand its industrial applications.
Below is a refined overview of its key features and implications:
Hardware: Trained on 120 NVIDIA H100 GPUs interconnected via NVIDIA Quantum-2 InfiniBand.
Efficiency: Completed training in 4 weeks using 2,688 GPU days, demonstrating cost-effective resource utilization.
Data: Utilized 98B tokens of high-quality Chinese pre-training data, enhanced through proprietary data-cleaning methods.
Architecture:
Context window: 128K tokens.
Adaptive Reasoning Reflection technology to improve autonomous reasoning.
Multi-node parallel training framework for stability
Performance (TMMLU+ Benchmarks):
Outperforms Taiwan’s Llama-3-Taiwan-70B in most domains, particularly mathematical and logical reasoning.
Matches global standards, nearing performance levels of top models like DeepSeek’s distillation variant.
Surpasses Meta Llama 3.1’s baseline in math-specific tasks
Comparison to DeepSeek’s V3
Judging from available public information, FoxBrain surpasses DeepSeek V3 in mathematical and logical reasoning due to its focus on structured problem-solving. DeepSeek V3, however, is better suited for creative writing and generic question-answering tasks.
In terms of speed and efficiency, DeepSeek V3 benefits from its Mixture-of-Experts architecture, allowing it to respond more quickly, whereas FoxBrain's efficiency lies in its optimized training process, which reduces computational costs.
The FoxBrain model, initially designed for internal applications by Hon Hai Research Institute, will be shared externally through open-source collaboration with technology partners. This move aims to expand FoxBrain's application scope, promoting AI adoption in manufacturing, supply chain management, and intelligent decision-making. During its development, NVIDIA provided crucial support with the Taipei-1 supercomputer and technical consulting, enabling the successful completion of the model training using NeMo. FoxBrain not only marks a significant breakthrough in Hon Hai's AI research but also sets a new benchmark for Taiwan's AI industry, poised to influence various scenarios and drive enterprise digital transformation and global industrial upgrading.
In the future, Hon Hai plans to leverage AI large language models like FoxBrain to optimize data analysis efficiency across key platforms: smart manufacturing, smart electric vehicles, and smart cities. By integrating FoxBrain into these sectors, it will become a pivotal engine for upgrading intelligent applications, ultimately enhancing the company's operational efficiency. This strategic integration underscores Hon Hai's commitment to harnessing AI for transformative growth, positioning FoxBrain as a cornerstone in driving innovation and competitiveness in the tech industry.
Besides HHRI, several Taiwanese institutions and companies are actively involved in developing Mandarin LLMs:
MiuLab: Known for the Taiwan LLM Initiative, MiuLab has developed models like Llama-3-Taiwan-70B, which is finetuned for traditional Mandarin and English users. This model demonstrates strong capabilities in language understanding and generation.
Project TAME: Led by a consortium including Chang Chun Group, Pegatron, and Unimicron, Project TAME aims to develop the world's first traditional Chinese language expert model. It promotes open-source development with support from Nvidia Taiwan.
National Science and Technology Council: They have developed TAIDE (Trustworthy AI Dialogue Engine), a generative AI chatbot optimized for traditional Chinese characters and Taiwanese cultural nuances. TAIDE is designed for tasks like article writing, translation, and summarization.
FoxBrain's launch marks a pivotal moment in Taiwan's AI journey, demonstrating the company's ability to develop cutting-edge, culturally aligned language models. With its focus on efficiency and reasoning capabilities, FoxBrain not only enhances Taiwan's AI landscape but also positions HHRI as a leader in localized AI solutions, capable of competing with global giants while preserving cultural and linguistic integrity.
Taiwanese institutions continue to innovate in the AI space, and models like FoxBrain, TAME, and TAIDE are poised to drive significant advancements in industries ranging from manufacturing to education. As the AI ecosystem in Taiwan continues to evolve, the potential for these models to transform industries and foster international collaboration remains vast.