o1

o1

o1 achieves 86% accuracy on Mathematics Olympiad benchmarks (vs. GPT-4o's 13%), offers PhD-level STEM proficiency, and maintains a 0.17% deceptive response rate in synthetic testing.

ConversationWeb DesignCode GenerationReasoning
Provider
Openai
Release Date
2024-12-05
Size
MEDIUM
Parameters
175B

Benchmark Performance

Performance metrics on industry standard AI benchmarks that measure capabilities across reasoning, knowledge, and specialized tasks.

MMLU

91.8%
View Source

GPQA

75.7%
View Source

MATH

96.4%
View Source

MGSM

89.3%
View Source

Model Insights

All Model Responses

Related Models