Announcing GPT-4o mini: Most cost-efficient small model

Logan K.
Logan K.StaffOpenAI StaffJul 18, 2024

We're excited to announce GPT-4o mini, our most cost-efficient small model. It's priced at 15 cents per million input tokens and 60 cents per million output tokens, making it significantly cheaper than previous models.

Key highlights

  • Scores 82% on MMLU, outperforming GPT-3.5 Turbo and comparable to GPT-4 on many benchmarks
  • 128K context window
  • Support for text, vision, and function calling
  • Available now in the API

To get started:

from openai import OpenAI
client = OpenAI()

response = client.chat.completions.create( model="gpt-4o-mini", messages=[{"role": "user", "content": "Hello!"}] )

Check the documentation for full details. Let us know your feedback!

14.2k views67 replies245 likes
6 Replies
Marcus Webb

This pricing is incredible. We can finally afford to run GPT-4 level models for our consumer app. Great work!

Sarah Chen

Benchmarks look solid. How does it perform on code generation compared to GPT-3.5 Turbo?

Logan K.

Great question! On HumanEval, GPT-4o mini scores 87.0% vs GPT-3.5 Turbo's 72.6%. Significant improvement for code tasks.

Alex Petrov

Already migrated our production system. Latency is comparable to 3.5-turbo but quality is noticeably better. The function calling is also more reliable.

Jordan Lee

Is there a way to try this on the free tier? I don't want to add a credit card just to test.

Maria Santos

Yes! GPT-4o mini is available on the free tier with rate limits. No credit card needed to get started.

Log in to reply to this topic.