VCBENCH

QUICK STATS

MODELS EVALUATED

20

HIGHEST PRECISION

Reasoned-Rule-Mining 87.5%

HIGHEST F0.5 SCORE

Verifiable-RL 36.6%

ABOUT VC BENCH

VCBench introduces the first standardized benchmark for founder-success prediction in venture capital.

Model Rankings
RankModelOrganization
Precision (%)
Recall (%)
F₀.₅ (%)
1
Verifiable-RL
Vela + Oxford42.623.636.6
2
Policy-Induction
Vela + Oxford41.020.234.0
3
Random-Rule-Forest
Vela + Oxford42.512.128.1
4
verifiable-reasoning
Vela + Oxford30.621.027.7
5
large-founder-model-v0
Vela + Oxford31.717.527.2
6
GPT-4o
OpenAI30.016.325.7
7
FinGPT-VC2
Columbia24.427.224.9
8
GPT-4o-mini
OpenAI31.511.123.0
9
FinGPT-VC1
Columbia21.824.222.2
10
o3
OpenAI43.27.421.5
11
Reasoned-Rule-Mining
Vela + Oxford87.55.021.0
12
Gemini-2.5-Pro
Google17.158.019.9
13
DeepSeek-Reasoner
DeepSeek31.86.918.4
14
Claude-3.5-Haiku-Latest
Anthropic15.846.418.2
15
GPT-5
OpenAI59.14.216.2
16
Gemini-2.5-Flash
Google12.568.414.9
17
DeepSeek-Chat
DeepSeek80.63.012.1
18
Tier-1 VCs
Humans23.05.210.7
19
Random Classifier
Baseline9.09.09.0
20
Y Combinator
Humans14.06.98.6

VC Bench evaluates AI models on venture capital functionality tasks.

Updated in real-time as new submissions are processed.