The gold standard for measuring AI scientific reasoning capability
GPQA Diamond (Graduate-level Google-Proof Q&A Diamond) is a benchmark that tests AI systems on expert-level questions in biology, chemistry, and physics. The "Diamond" subset contains the most challenging questions requiring deep scientific reasoning that cannot be answered through web search.
| Rank | Model | Accuracy | Type |
|---|---|---|---|
| 1 | Omic AI Scientist | 93.3% | Specialized (Bio/Chem) |
| 2 | Gemini 3 Pro | 91.9% | General-purpose LLM |
| 3 | GPT-5.1 | 88.1% | General-purpose LLM |
| 4 | Claude 4.5 Sonnet | 83.4% | General-purpose LLM |
| — | Human PhD Expert | 69.7% | Human baseline |
Drug discovery requires expert-level reasoning in biology, chemistry, and pharmacology. GPQA Diamond measures exactly this capability. AI systems that score highly on GPQA Diamond can:
Reason about molecular mechanisms, pathway dysregulation, and disease etiology at PhD level.
Understand protein-ligand binding, enzyme kinetics, and chemical reactivity.
Interpret complex biological datasets and draw valid scientific conclusions.
Propose rational drug design approaches based on mechanistic understanding.
Omic's AI Scientist achieves 93.3% on GPQA Diamond—the highest score among all tested systems. This performance comes from:
Focused on biology, chemistry, and drug discovery rather than general knowledge.
Deep understanding of genomics, proteomics, and metabolomics relationships.
Trained on disease mechanisms, not just isolated facts.
Human PhD experts score around 69.7%. Scores above 70% indicate PhD-level performance; above 90% indicates superhuman scientific reasoning. Top AI systems now exceed 90%.
As of November 2025, Omic AI Scientist leads with 93.3% accuracy, followed by Gemini 3 Pro (91.9%) and GPT-5.1 (88.1%).
Unlike general knowledge benchmarks, GPQA specifically tests graduate-level scientific reasoning. Questions cannot be answered through memorization or web search, making it relevant for evaluating AI for scientific research.