VCBENCH

QUICK STATS

MODELS EVALUATED

22

HIGHEST PRECISION

Reasoned-Rule-Mining 87.5%

HIGHEST F0.5 SCORE

Verifiable-RL 36.6%

ABOUT VC BENCH

VCBench introduces the first standardized benchmark for founder-success prediction in venture capital.

Model Rankings
RankModelOrganization
Precision (%)
Recall (%)
F₀.₅ (%)
1
Verifiable-RL
Vela + Oxford42.623.636.6
2
Policy-Induction
Vela + Oxford41.020.234.0
3
GemVC-v0
Independent39.420.332.9
4
Structured-Rule-Stump
Independent32.818.028.1
5
Random-Rule-Forest
Vela + Oxford42.512.128.1
6
verifiable-reasoning
Vela + Oxford30.621.027.7
7
large-founder-model-v0
Vela + Oxford31.717.527.2
8
GPT-4o
OpenAI30.016.325.7
9
FinGPT-VC2
Columbia24.427.224.9
10
GPT-4o-mini
OpenAI31.511.123.0
11
FinGPT-VC1
Columbia21.824.222.2
12
o3
OpenAI43.27.421.5
13
Reasoned-Rule-Mining
Vela + Oxford87.55.021.0
14
Gemini-2.5-Pro
Google17.158.019.9
15
DeepSeek-Reasoner
DeepSeek31.86.918.4
16
Claude-3.5-Haiku-Latest
Anthropic15.846.418.2
17
GPT-5
OpenAI59.14.216.2
18
Gemini-2.5-Flash
Google12.568.414.9
19
DeepSeek-Chat
DeepSeek80.63.012.1
20
Tier-1 VCs
Humans23.05.210.7
21
Random Classifier
Baseline9.09.09.0
22
Y Combinator
Humans14.06.98.6

VC Bench evaluates AI models on venture capital functionality tasks.

Updated in real-time as new submissions are processed.