GSM8K
GSM8K, short for Grade School Math 8K, is a benchmark introduced by OpenAI in 2021 containing 8,500 grade-school-level multi-step math word problems requiring 2-8 steps of reasoning to solve. The benchmark became iconic in
LLM evaluation because solving GSM8K reliably requires the kind of multi-step reasoning that smaller models struggle with, making it a litmus test for emergent reasoning capabilities. Chain-of-thought prompting was first demonstrated dramatically on GSM8K, where step-by-step prompting improved
PaLM's accuracy from 18% to 57%. Reference scores include GPT-3 (8.6%), Llama 2 70B (56.8%), GPT-4 (92%), Claude 3 Opus (95.0%), Claude 3.5 Sonnet (96.4%), o1-preview (94.8%), and o1 (96.4%). GSM8K is now considered essentially saturated by frontier models, and harder math benchmarks like MATH, AIME, FrontierMath, and Putnam have taken over as discriminating evaluations. AI governance teams use GSM8K for baseline reasoning capability validation but rely on harder benchmarks for current model comparison. The dataset is hosted on Hugging Face and the original GitHub repository.
Math-reasoning models with Centralpoint: Centralpoint routes mathematical reasoning workloads to models validated on GSM8K, MATH, and AIME — Claude 3.5 Sonnet, o1, Gemini 2.5, DeepSeek-R1 — in a model-agnostic stack. Tokens are metered per skill, prompts stay local, and reasoning-aware chatbots deploy through one line of JavaScript on any portal.
Related Keywords:
GSM8K,
,