v1-beta · January 2026

LMIQ

A new benchmark for frontier AI
Intelligence is not one-dimensional. AGI will be multi-dimensional. New research is needed.

Methodology

LMIQ is composed of a diverse set of novel challenges which are designed to leverage multi-dimensional capabilities of human intelligence. The gap between human performance and leading AI systems remains significant.

Frontier AI models remain

~10%
As accurate
Average humans score 95-100%
5x
Less Time-Efficient
Human reasoning is significantly faster
100x
Less Energy-Efficient
Biological compute is more energy-efficient

Benchmark Results

These are preliminary benchmark beta results with wide confidence intervals. AI scores were produced on test sets composed of 50 challenges, ranging in difficulty. Humans were tested on the same challenge set.

Scaling alone will not produce AGI.
New research is needed.