o1-preview

OpenAI's reasoning model trained to think before responding, achieving PhD-level science performance and strong results in math and competitive coding.

about

The first commercial model to surpass PhD-level human experts on GPQA Diamond, o1-preview scored 77.3% against the 69.7% achieved by domain specialists with internet access. It introduced inference-time compute scaling, where spending more tokens on hidden chain-of-thought reasoning produces measurably better answers, placing among the top 500 US students on AIME and in the 89th percentile on Codeforces.

Licenseopenai
Context window(in thousands)128,000

Use cases for o1-preview

  1. PhD-level scientific reasoning: Scoring 77.3% on GPQA Diamond, 8 points above human PhD experts with internet access, o1-preview handles graduate-level physics, chemistry, and biology problems requiring multi-step inference.
  2. Competitive programming: Placing in the 89th percentile on Codeforces, it solves algorithmic challenges that require sustained logical reasoning, constraint analysis, and optimal solution design.
  3. Inference-time compute scaling: Its hidden chain-of-thought architecture produces measurably better answers when given more thinking tokens, enabling developers to trade latency for accuracy on high-stakes queries.

Quality

Arena Elo1388
MMLUN/A
MT BenchN/A

o1-preview scores 90.8% on MMLU (0-shot CoT) and 78.0% on GPQA Diamond, surpassing human PhD experts (69.7%) on the latter by 8 points. Compared to GPT-4 Turbo (86.5% MMLU) on the same sheet, the 4-point MMLU improvement understates the reasoning leap, as o1-preview's hidden chain-of-thought architecture specifically targets multi-step inference that standard benchmarks underweight.

MiniMax-M2.5

1406

gpt-5-mini

1390

o1-preview

1388

gpt-4.1-mini

1382

gpt-4o-mini

1382

pricing

Running o1-preview through Telnyx Inference costs $15.00 per million input tokens and $60.00 per million output tokens. Processing 100,000 PhD-level science problems at 5,000 tokens each would cost approximately $37,500. For STEM-only workloads, o1-mini ($3/$12 per million tokens) on the same sheet delivers 80% of the reasoning at 80% lower cost.

What's Twitter saying?

  • Developers praise o1-preview for superior coding quality, producing fewer bugs and solving complex problems that other models couldn't, though it often generates verbose outputs and unexpectedly refactors code.
  • Benchmarks and tests show strong reasoning and math performance, but it's only okay at coding—often outperformed by Claude 3.5 Sonnet—and weaker in creative writing.
  • Many note o1-preview outperformed the newer o1 model in creative problem-solving and autonomy, with o1 seen as lazier, more rigid, and a regression.

Explore Our LLM Library

Discover the power and diversity of large language models available with Telnyx. Explore the options below to find the perfect model for your project.

Organizationdeepseek-ai
Model NameDeepSeek-R1-Distill-Qwen-14B
Taskstext generation
Languages SupportedEnglish
Context Length43,000
Parameters14.8B
Model Tiermedium
Licensedeepseek

TRY IT OUT

Chat with an LLM

Powered by our own GPU infrastructure, select a large language model, add a prompt, and chat away. For unlimited chats, sign up for a free account on our Mission Control Portal here.

HOW IT WORKS

Selecting LLMs for Voice AI

RESOURCES

Get started

Check out our helpful tools to help get you started.

  • Icon Resources ebook

    Test in the portal

    Easily browse and select your preferred model in the AI Playground.

  • Icon Resources Docs

    Explore the docs

    Don’t wait to scale, start today with our public API endpoints.

  • Icon Resources Article

    Stay up to date

    Keep an eye on our AI changelog so you don't miss a beat.

Sign up and start building

faqs

Is o1 Preview better than o1?

o1 Preview was the early access version of OpenAI's o1 reasoning model, later superseded by the full o1 release. The full o1 offers improved accuracy and reliability over the preview version.

What is o1 Preview good for?

o1 Preview excels at STEM reasoning, mathematical problem-solving, and complex analytical tasks. It uses internal chain-of-thought reasoning before responding, making it suited for technical applications that benefit from deeper thinking.

Is o1 Preview available?

o1 Preview remains accessible through the API but has been superseded by the full o1 and o3-mini models. OpenAI recommends migrating to newer reasoning models for new projects.

How much does o1 Preview cost?

o1 Preview is priced at $15 per million input tokens and $60 per million output tokens, reflecting the additional compute used for internal reasoning. The newer o3-mini offers similar reasoning at lower cost.

Is o1 Preview better than GPT-4?

o1 Preview significantly outperforms GPT-4 on mathematical reasoning, science, and coding tasks that require step-by-step analysis. GPT-4 remains stronger for general-purpose tasks where speed matters more than maximum reasoning depth.

What is the difference between o1 Preview and o1 mini?

o1 Preview is the full reasoning model with broad capabilities, while o1 mini is a smaller, cheaper variant optimized for STEM tasks. o1 mini offers 80% cost savings with minimal accuracy loss on math and coding.