Taiwan's tech giant Foxconn has just stepped into the AI spotlight with the launch of its very first large language model, and it's got big plans for it. Named "FoxBrain," the model is set to power smarter manufacturing and streamline supply chain operations—a major move for the world's largest contract electronics maker.
Foxconn, known for assembling Apple's iPhones and producing Nvidia's AI servers, revealed that FoxBrain is built on Meta's Llama 3.1 architecture. The company trained the model using 120 of Nvidia's powerful H100 GPUs, wrapping up the process in just four weeks.
What makes FoxBrain stand out is that it's the first of its kind in Taiwan with advanced reasoning capabilities, and it's specifically tuned for traditional Chinese and Taiwanese language use. While Foxconn acknowledged that it still slightly trails behind China's DeepSeek distillation model in terms of performance, it's already brushing shoulders with world-class AI standards.
Initially, FoxBrain will be used internally at Foxconn, helping with everything from data crunching and decision-making to document collaboration, math, reasoning tasks, and even generating code. But the company doesn't plan to keep it all in-house. They're looking to partner with other tech players, share some of the model's open-source data, and promote the use of AI in manufacturing, logistics, and intelligent decision-making across the board.
Foxconn also gave a nod to Nvidia for its support, especially with access to "Taipei-1"—Nvidia's massive supercomputer based in Kaohsiung, which played a key role in training FoxBrain.
More on FoxBrain will be unveiled at Nvidia's GTC developer conference in mid-March, so stay tuned for deeper insights into what could be Taiwan's next big leap in AI.
Comments