Find the weights that matter

In a 70-billion parameter model, protecting just 50 weights
during 4-bit quantization recovers 90-99% of lost quality.

See the resultsData-free. Single forward pass.
50
Weights Protected
99%
Quality Recovery
1
Forward Pass
0
Calibration Data

How PeakWeights works

A simple scoring function finds the needles in 70 billion haystacks.

1

Score Every Weight

score = |weight| × |max_activation|

Captures worst-case quantization error.

2

Find Top-K

In 70 billion parameters, only ~50 are critical.

They follow a power law distribution.

3

Protect & Quantize

Keep critical weights in FP16, quantize the rest to 4-bit.

Recover 90-99% of lost quality.

Proven results

Tested across 5 architectures from the research paper

ModelFP16 PPL4-bit PPL+PeakWeightsRecovery
Qwen2.5-7B10.6211.5110.6399%
Mistral-7B-v0.313.4413.813.5861%
SmolLM2-1.7B17.9224.5617.9999%
DeepSeek-R1-Distill-Qwen-7B70.7373.5270.9193%
Phi-3-mini11.6512.6711.6897%

Perplexity on WikiText-103. Lower is better.

Quick start

Install and run in under a minute

1. Install

pip install peakweights

2. Find critical weights

peakweights Qwen/Qwen2.5-7B --top_k 50

3. Or use Python

from peakweights import find
critical = find("Qwen/Qwen2.5-7B", k=50)

Top 25 models database

Pre-computed peak weights for the most popular open-source LLMs in 2026

DeepSeek
DeepSeek
DeepSeek
DeepSeek
Qwen3
Alibaba
Llama 4 Maverick
Meta
Llama 4 Scout
Meta
Qwen2.5
Alibaba
Llama
Meta
DeepSeek
DeepSeek
Mixtral
Mistral AI
Qwen2.5
Alibaba

Read the paper

“PeakWeights: Data-Free Discovery of Critical LLM Parameters”

Thiyagarajan M (Kalmantic Labs) & Vamshi Ambati (CMU, IIIT Hyderabad)

Read Paper