Foxconn Unveils First Massive Language Mannequin Dubbed ‘FoxBrain’

Taiwan’s Foxconn mentioned on Monday it has launched its first giant language mannequin and plans to make use of the know-how to enhance manufacturing and provide chain administration.
The mannequin, named “FoxBrain,” was skilled utilizing 120 of Nvidia’s H100 GPUs and accomplished in about 4 weeks, the world’s largest contract electronics producer mentioned in an announcement.
The corporate, which assembles iPhones for Apple and likewise produces Nvidia’s synthetic intelligence servers, mentioned the mannequin relies on Meta’s Llama 3.1 structure.
It’s Taiwan’s first giant language mannequin with reasoning capabilities that’s optimised for conventional Chinese language and Taiwanese language kinds, it mentioned.
Foxconn mentioned that although there was a slight efficiency hole in contrast with China’s DeepSeek’s distillation mannequin, its general efficiency may be very near world-class requirements.
Initially designed for inside purposes, FoxBrain covers information evaluation, resolution assist, doc collaboration, arithmetic, reasoning and problem-solving, and code technology.
Foxconn mentioned it plans to collaborate with know-how companions to broaden the mannequin’s purposes, share its open-source data, and promote AI in manufacturing, provide chain administration, and clever decision-making.
Nvidia supplied assist via its Taiwan-based supercomputer “Taipei-1” and provided technical consulting throughout the mannequin’s coaching, Foxconn mentioned.
Taipei-1, the biggest supercomputer in Taiwan, is owned and operated by Nvidia in Kaohsiung, a southern metropolis on the island.
Foxconn will announce additional particulars concerning the mannequin throughout Nvidia’s GTC developer convention in mid-March.
© Thomson Reuters 2025
(This story has not been edited by NDTV workers and is auto-generated from a syndicated feed.)