Skip to main content
Back to Pulse
Crescendo AI

Meta announces four new in-house AI chips

Read the full articleMeta Announces Four New In-House AI Chips on Crescendo AI

What Happened

Meta announced the MTIA 300-500 series, custom AI accelerators designed to reduce dependency on NVIDIA GPUs for its inference and training workloads. The company plans mass deployment by end of 2027. Meta joins Google (TPUs), Amazon (Trainium), and Microsoft (Maia) in pursuing custom silicon to manage compute costs at scale.

Our Take

Look, every major tech company building custom silicon isn't news anymore — Google's had TPUs for a decade. What's different here is the scale: Meta's burning through H100s at a rate that apparently makes it worth shipping their own hardware. That's a serious pain point.

MTIA 300-500 series. The real question nobody's asking is whether these are inference chips or training chips — because that determines whether this eventually matters to us as API consumers or stays internal.

Here's the thing about compute pricing: it doesn't drop because someone announces competing silicon. It drops when that silicon ships at scale and the hyperscalers actually pass savings downstream. Mass deployment by end of 2027 means we're 18+ months out from any real pricing pressure on APIs we use.

Honestly? NVIDIA's stock dipped on the news, but they're not sweating. One customer vertically integrating doesn't kill a monopoly. But four major customers all cutting custom chips simultaneously? That's a different conversation over a 5-year horizon.

Watch Llama API pricing in late 2027. That's the actual signal — not the chip announcement.

What To Do

Log your current per-token costs on Llama-based APIs (together.ai, Fireworks, Meta's own API) today — you'll want a baseline to measure against when MTIA production chips supposedly hit in 2027.

Builder's Brief

Who

AI infrastructure engineers benchmarking alternatives to NVIDIA for inference at scale

What changes

Meta's inference cost trajectory and latency SLAs may shift, affecting third-party applications built on Meta AI APIs

When

months

Watch for

MTIA 300 benchmark results against H100 and H200 on Meta's own production model serving workloads

What Skeptics Say

Meta has announced custom silicon roadmaps before without materially reducing NVIDIA dependence; a mass deployment target of end-2027 gives NVIDIA two full product generations to widen the performance gap before Meta ships at scale.

Cited By

React

Newsletter

Get the weekly AI digest

The stories that matter, with a builder's perspective. Every Thursday.

Loading comments...