ARTFEED — Contemporary Art Intelligence

New Benchmark Estimates LLM Parameter Counts via Factual Knowledge

ai-technology · 2026-04-30

Researchers have developed a novel technique to estimate the parameter counts of closed-source large language models by evaluating their factual knowledge. They created Incompressible Knowledge Probes (IKPs), consisting of 1,400 factual questions categorized into 7 levels of obscurity, aimed at identifying knowledge that cannot be inferred through reasoning or improved by architectural changes. By establishing a log-linear relationship between IKP accuracy and parameter count using 89 open-weight models (ranging from 135M to 1,600B parameters) from 19 different vendors, they achieved an R² value of 0.917. The leave-one-out cross-validation revealed a median fold error of 1.59×, with 68.5% of estimates falling within 2× and 87.6% within 4×. This method leverages the fundamental requirement that storing F facts necessitates at least F/(bits per parameter) weights, offering a more precise estimate than inference economics, which incurs over 2× uncertainty due to hardware, batching, and serving-stack assumptions. The research can be found on arXiv under ID 2604.24827.

Key facts

  • Method estimates LLM parameter counts via factual knowledge
  • IKP benchmark has 1,400 factual questions across 7 obscurity tiers
  • Calibrated on 89 open-weight models from 19 vendors
  • Model sizes range from 135M to 1,600B parameters
  • R² of 0.917 for log-linear mapping
  • Median fold error 1.59× in cross-validation
  • 68.5% of estimates within 2×, 87.6% within 4×
  • Paper ID: arXiv:2604.24827

Entities

Institutions

  • arXiv

Sources