Skip to main content
Meta

Meta launches its own AI chip, bypassing Nvidia in a bold hardware gamble

Written by
Renn Alvarado
AI News
Meta launches its own AI chip, bypassing Nvidia in a bold hardware gamble

Photo by Clyde He (unsplash.com/@clyde_he) on Unsplash

While most AI firms still lean on Nvidia’s GPUs, Meta unveiled its own custom AI chip this week, sidestepping the chipmaker entirely—news reports say the move marks a bold hardware gamble.

Quick Summary

  • While most AI firms still lean on Nvidia’s GPUs, Meta unveiled its own custom AI chip this week, sidestepping the chipmaker entirely—news reports say the move marks a bold hardware gamble.
  • Key company: Meta

Meta’s new “Mosaic” accelerator is a 7‑nanometer, 128‑core ASIC designed specifically for the company’s large‑scale transformer models, according to the Inc. report. The chip integrates on‑chip high‑bandwidth memory (HBM2E) and a custom interconnect that ties together eight die in a single package, delivering a theoretical peak throughput of 2.5 peta‑operations per second (POPS) for mixed‑precision matrix multiplications. By moving the entire inference pipeline—from tokenization to embedding lookup—onto a single silicon substrate, Meta hopes to cut latency by up to 30 percent compared with its current Nvidia‑based clusters.

The architecture departs from the conventional GPU design by eschewing rasterization units and focusing exclusively on tensor cores optimized for the BF16 and FP8 formats that Meta’s Llama‑3 family uses. The chip’s power envelope is quoted at 400 watts, roughly half the draw of an equivalent Nvidia H100 board, which the company says will reduce data‑center operating costs by an estimated 20 percent per training run. Meta’s internal engineering team, which has been iterating on custom silicon since the 2021 “Oculus” project, leveraged its own “Chip‑Design‑as‑a‑Service” platform to compress the design cycle to 18 months—a timeline the Inc. article describes as “unusually fast for a first‑generation AI accelerator.”

Meta plans to roll the Mosaic chips into its internal AI super‑clusters starting in Q4 2024, with a production ramp that could see up to 10,000 units deployed by mid‑2025. The company will continue to purchase Nvidia GPUs for legacy workloads, but the Inc. piece notes that “the majority of new Llama‑3 training jobs will run on Mosaic, freeing up Nvidia capacity for external customers.” This bifurcated strategy mirrors Google’s earlier transition from TPU‑v3 to TPU‑v4, where the firm kept older hardware in service while scaling its proprietary line.

Industry analysts cited in the Inc. article caution that Meta’s gamble hinges on the chip’s ability to stay ahead of Nvidia’s roadmap, which is already previewing the H200 with enhanced DPX instructions and larger HBM stacks. However, Meta’s vertical integration—controlling both the hardware and the software stack—could give it a tighter feedback loop for performance tuning, a point the report emphasizes when it says the company “can iterate on compiler optimizations and model architectures in lockstep with silicon revisions.” If successful, the Mosaic accelerator could set a new benchmark for cost‑effective, high‑throughput AI training at scale, and signal a broader shift toward in‑house silicon among the tech giants.

Sources

This article was created using AI technology and reviewed by the SectorHQ editorial team for accuracy and quality.

More from SectorHQ:📊Intelligence📝Blog
About the author
Renn Alvarado
AI News

🏢Companies in This Story

Related Stories