lfm2-24b-a2b

THINKING MODEL

Mac mini (Apple M4 Pro)

64 GB · macOS 15.7.4

Tested on March 6, 2026
Top 100% Compare
Global Score
18 /100
Not Rec.
Hardware Fit
59/100
Quality
0/100

Get this model

Hardware

Machine
Mac mini
CPU
Apple M4 Pro
Cores
14 total (10 perf + 4 eff)
Frequency
2.4 GHz
RAM
64 GB LPDDR5
GPU
Apple M4 Pro
OS
macOS 15.7.4
Arch
arm64
Power Mode
balanced

Performance

Tokens/sec
6.6
Standard deviation
±4.3
First chunk latency
1.0 s
Time to first token
1.0 s
Load time
N/A
Memory usage
0.0 GB (0%)
Total tokens
558

Score breakdown

Speed
9/50
Time to first token
20/20
Memory
30/30

Quality

Reasoning
0/20
Coding
0/20
Instruction following
0/20
Structured output
0/15
Math
0/15
Multilingual
0/10

Category levels

Reasoning: Poor Coding: Poor Instruction Following: Poor Structured Output: Poor Math: Poor Multilingual: Poor

Metadata

Spec version
0.2.1
Runtime
LM Studio 0.4.6+1
Model format
GGUF
Hardware profile
HIGH-END
Result hash
ee428d98cb597d14e3f6228b34d1587ca2929ee82106ee04e1d3c90647797460

Interpretation

Hardware fit: 59/100. Overall suitability: NOT RECOMMENDED (Global 18/100). Category profile: Reasoning: Poor, Coding: Poor, Instruction Following: Poor, Structured Output: Poor, Math: Poor, Multilingual: Poor. Warning: model produced very low accuracy on quality tasks — results may be unusable despite good hardware performance.

Warnings

  • Model produced very low accuracy on quality tasks — results may be unusable despite good hardware performance.
  • Token speed is unstable (stddev 4.3 tok/s, mean 6.6 tok/s) — may indicate thermal throttling or memory pressure.

Bench Environment

Power: AC CPU load: avg 29% (peak 34%)

Run yours now

$ npm install -g metrillm@latest
$ metrillm

Requires Node 20+ and Ollama or LM Studio running

Or run without installing: npx metrillm@latest