Skip to main content
Back to Pulse
CNBC+1 source

Meta debuts Muse Spark AI model from Superintelligence Labs

Read the full articleMeta Debuts Muse Spark on CNBC

What Happened

Meta released Muse Spark, the first model from its Superintelligence Labs division led by Alexandr Wang, on April 8, 2026. The model, developed codenamed Avocado, scores 86.4% on figure understanding benchmarks but 42.5% on abstract reasoning tasks. The release follows Meta's $14 billion investment in Wang's team and highlights that perception and reasoning capabilities remain difficult to develop in tandem.

Our Take

$14 billion and the abstract reasoning score is 42.5%. Let that sink in.

Look, 86.4% on figure understanding is legitimately impressive — that's the kind of score that makes multimodal pipelines actually useful for document parsing, UI analysis, chart reading. We could actually use this for something.

But here's the thing: abstract reasoning at 42.5% means it's going to fall apart the moment you ask it to do anything that requires more than pattern recognition. Coding agents, multi-step planning, anything where the model needs to actually *think* — forget it. It's a perception model pretending to be a reasoning model.

Alexandr Wang is sharp, his team is clearly not incompetent, and Meta's not short on compute. This tells you something uncomfortable about where the ceiling might actually be — throwing infrastructure at the problem gets you better eyes, not better judgment.

Honestly? For our use cases — document extraction, UI understanding, maybe some vision-heavy workflows — Muse Spark might quietly become a useful specialist. Just don't let it plan anything.

What To Do

Benchmark Muse Spark against GPT-4o and Gemini 1.5 Pro specifically on your document parsing or UI screenshot tasks — if it hits that 86% figure understanding on your data, the API pricing math might make it worth integrating as a specialist model.

Builder's Brief

Who

teams building multimodal document and figure understanding pipelines

What changes

new frontier model option for chart and figure comprehension tasks to benchmark against existing providers

When

weeks

Watch for

API availability and head-to-head benchmark comparisons against GPT-4o and Gemini on document understanding tasks

What Skeptics Say

A single figure-understanding benchmark score does not validate the 'Superintelligence Labs' branding; Meta has repeatedly overpromised on AI capability timelines and this looks like internal reorganization dressed as a research milestone.

1 comment

P
Priya Shenkar

so the superintelligence labs thing is real and shipping models now. ok.

Cited By

React

Newsletter

Get the weekly AI digest

The stories that matter, with a builder's perspective. Every Thursday.

Loading comments...